Thanks to visit codestin.com
Credit goes to github.com

Skip to content

Tags: kaitz/fxcm

Tags

v26

Toggle v26's commit message
Update

Known dictionary words are compared with their codeword. Previously, text strings were compared.
Adjusted global StateMap prediction.
ContextMap (HT 128) reduced predictions from 6/4 to 4/3 per context. Use single internal StateMap. All context states are update with that.
ContextMap (HT 32) reduced predictions from 5/4 to 3/2 per context. Removed StateMap based predictions.
StationaryMap for 2 context
In WordsContext use also codeword for dictionary word.
Added SentenceContext for sentance managment. Max 64 sentances (WordsContexts). Search for similarity is performed by compareing codewords (default 53% means match found).
In stemmer add Pronoun word type.
Add InDirectStateMap with order-w mixing of primary predictions (similar to Paq9a/zpaq)
Partial sentance contexts.
Group of SentenceContexts for: lists ('*'), table, wikilinks and regular sentances. Total 4.
Removed SparseMatchModel.
Removed 4 SmallStationaryContextMap contexts
Mixer count from 12 to 24
Added 7 new ContextMap's
Added 22 new InDirectStateMap contexts
Adjusted mixer parameters and contexts
Adjusted ContextMap memory usage
There are 3 mixers layers (+1 in every InDirectStateMap).
For layer 0 mixers about ~40% of updates are skipped.
Some predictions are skipped if line is Category link, after topic 'See also', 'References', 'Bibliography' or 'External links'.
Some low memory ContextMaps are reset after every page (wikipedia article). The StateMap is preserved if it exists.

v24

Toggle v24's commit message
Update fxcm.cpp

* Update model to fx2-cmix level
* Add match skip

v22

Toggle v22's commit message
v22

* Reverse dictionary transform. We load the dictionary when it is found after decompressing it. Text has a separate buffer from coded byte stream buffer.
* Natural language processing using stemmer (from paq8px(d)).
* Stemmer has new word types: Article, Conjunction, Adposition, ConjunctiveAdverb.
* Some word (related) contexts are changed based on what type of word was last. Some words are removed from word streams depending on the last word type. This improves compression.

There are four word streams:
* 1. basic stream of undecoded words.
* 2. decoded word stream after stemming for sentences. Contains all words. Reset when sentence ends.
* 3. decoded word stream after stemming for paragraphs. Contains words that are not: Conjunction, Article, Male, Female, ConjunctiveAdverb. Reset when paragraph ends.
* 4. decoded word stream after stemming. Contains words that are not: Conjunction, Article, Male, Female, Adposition, AdverbOfManner, ConjunctiveAdverb.
* Word limit per stream is increased from 64 to 256 words.
* New context that uses stemmer and decoded plaintext. Some global context are changed when word type is: \
ConjunctiveAdverb or Conjunction - skip updating in stream 1. Conjunction for sentence reset. etc. Knowing these new words allowed large amounts of compression improvements.
* In some cases words are removed between certain chars from stream 2 and 3 when following is true:  \
=| - wiki template \
<> - html/xml tags \
[| - wiki links \
() - usually words in sentences
* Main predictors are split between three different ContextMaps. This provides better compression. Sizes for hash tables are 32, 64 (standard for paq8 versions) or 128 bytes per contexts. 32 byte size is good for small memory context (below 256 KB), 64 is good for medium sized context (up-to 16MB), 128 is good for large memory context (more than 16MB).
* One state table is removed and replaced with another one. State tables are generated at runtime to reduce code size.
* Added sparse match model. With a gap of 1-2 bytes and minimum length of 3-6 bytes. Mostly for escaped UTF8.
* Detection of math, pre, nowiki, text tags in decoded text stream. Some word related contexts are not used when 3 first tags content is compressed. Improves compression speed.
* More parsing of lists and paragraphs. So that context for predictors is best as they can.
* Optimized context skipping in main predictors.
* Main predictor context bias is not forwarded to cmix floating point mixers, instead a single prediction bias is set. This avoids unnecessary expansion of mixer weight space and maintains lower memory/cpu usage. There are some other predictions that are not forwarded as they make compression worse and slower.
* Some mixers and APM’s context size is larger so that prediction can be better.
* partially/fully decoded word index into dictionary is used as a context for mixer in fxcm mixer
* Some variables are renamed for better readability.

v17

Toggle v17's commit message
v17-v18

Some cleanup
Adjust contexts

v15

Toggle v15's commit message
v15-v16

-Add one new context
-Adujust on mixer
-Some cleanup
-Add comments so reader can understand what is happening
-Tune some variables and contexts
-Improve compression speed

v13

Toggle v13's commit message
v13-v14

-Word&Sentence context.
-Adjust some contexts

v11

Toggle v11's commit message
v11-v12

-Wiki table/row & column context
-Add 3 new context based on table
-Change some context

v9

Toggle v9's commit message
v9-v10

-Change some contexts
-Increase one mixer size

v7

Toggle v7's commit message
Page

One article as seen by first char context

v5

Toggle v5's commit message
v5-6

-Add another context based on first char
-Adjust some contexts