Now instead of create a context with the number of file to parse and
check every x millisecond if it is complete, I use work group. For each
file, I spawn a thread in the group and then I call wg.wait()
Got an error that I could find for a while. It was when parsing a lot of
files.
Turn out it was the Thread Pool or something because if I run on 1 core,
it's ok.
Different thing in this commit, I just want to freeze a state that
benchmark work.
Before I was parsing all files, getting the max index then parsing from
0 to the max.
Byt now that I delete empty files, I need to parse only existing one.
Added NOW already and now debuging some stuff regarding filter and
parsing file of one struct when it should be another
Also moved query test into a seperated test file.
And some fix and changed in docs
Now I flush only when the file is full and I check the the currently
used file if it is big enough.
So I dont get stat of all files and flush everytime like before
It is working !
The basic GRAB [name, best_friend] work and do return the best friend
infos.
I added more test and some do not work yet, working on it now
This dosnt work yet but I implemented the reparsing of the files to
return relationship.
So GRAB User [name, friends] should return all infos of all friends and
not just the UUID of those friends like before
So then I can parse the file again, create a map of UUID sub json and
just iterate over the first json looking for UUID and update with new
data from map