Analysis: Difference between revisions

From SuperMemo Help
Jump to navigation Jump to search
(→‎Use statistics: Added option Tasklist (name of the tasklist))
(123 intermediate revisions by 3 users not shown)
Line 1: Line 1:
__TOC__
__TOC__


The tabbed dialog available with '''[[Tools menu|Tools]] : Statistics : Analysis''' (''Shift+Alt+A'') in SuperMemo provides graphs and tables that illustrate the current state of the learning process in the currently opened collection.  
The tabbed dialog available with '''[[Toolkit menu|Toolkit]] : [[Toolkit menu#Statistics|Statistics]] : Analysis''' (''Shift+Alt+A'') in SuperMemo provides graphs and tables that illustrate the current state of the learning process in the currently opened [[Glossary:Collection|collection]].  


In addition to '''Analysis''', SuperMemo provides multiple other [[Theory|analytical tools]].
In addition to '''Analysis''', SuperMemo provides multiple other [[Theory|analytical tools]].


== Distributions ==
== Use ==


=== Intervals ===
Graphs of changes in the statistics of the learning process over time. These graphs record daily changes to statistics such as the number of [[Glossary:Memorized_element|memorized]] [[Glossary:Item|items]], [[Glossary:Measured_forgetting_index|measured forgetting index]] or the number of [[Glossary:Outstanding_element|outstanding]] [[Glossary:Item|items]] for a given day. The first non-zero value of the day is recorded. If that value increases afterwards, that increase is ignored (except a few cases such as the statistic of newly [[Glossary:Memorized_element|memorized]] [[Glossary:Item|items]], etc.). If you have a break in learning, the last recorded non-zero value will propagate through the period in which you did not use SuperMemo (except a few cases such as the statistics of repetitions made, etc.).


'''Intervals''' - distribution of inter-repetition intervals in a given collection. By changing the setting of the combo box at the bottom of the dialog, you can see interval distributions for all elements in the collection, for only items, and for only topics.
To zoom in on a portion of the graph (as in pictures below), sweep the portions of the graphs that are to be removed with the mouse, i.e. point at the day of focus, move the mouse left and release the mouse button. This cuts away part of the graph in the direction that the mouse was moved. You can also use ''Ctrl+Left'' and ''Ctrl+Right'' buttons to zoom in (see [[#Use context menu | below]])


=== A-Factors ===
=== Use statistics ===


'''A-Factors''' - distribution of [[Glossary#A-factor|A-Factors]] in a given collection. The distribution itself is not used in [[SuperMemo Algorithm|Algorithm SM-15], and merely results from it. Note that topics and items use different A-Factor categories. Choose ''Items only'' or ''Topics only'' at the bottom to see distributions for items or topics. If you choose ''All elements'' you will see a less meaningful conglomerate distribution indexed by A-Factor category number.
Here are the statistics recorded daily by SuperMemo and displayed on the '''Use''' tab:  
[[Image:AFactors distribution.jpg|center|800px|SuperMemo: Tools : Statistics : Analysis : Distributions : A-Factors shows you how A-Factors are distributed in your collection]]


=== Repetitions ===
#'''Work done'''
#*'''Use time''' - your daily total time in which you actively used a given [[Glossary:Collection|collection]]. The time is taken from learning blocks record ('''[[Toolkit menu|Toolkit]] : [[Sleep Chart]]''' in red). Days count from midnight to midnight (i.e. midnight shift is disregarded)
#*'''Recall time''' - your daily total time which you spent on trying to recall answers to questions
#*'''Repetitions''' - your daily recorded number of [[Glossary:Item|item]] and [[Glossary:Topic|topic]] reviews
#*'''Item repetitions''' - your daily record of [[Glossary:Item|item]] repetitions
#*'''Topic reviews''' - your daily record of [[Glossary:Topic|topic]] reviews
#'''Outstanding''' - number of [[Glossary:Outstanding_element|outstanding elements]] scheduled for a given day
#*'''Outstanding''' - number of all [[Glossary:Outstanding_element|outstanding elements]] scheduled for a given day
#*'''Outstanding items''' - number of [[Glossary:Outstanding_element|outstanding]] [[Glossary:Item|items]] scheduled for a given day
#*'''Outstanding topics''' - number of [[Glossary:Outstanding_element|outstanding]] [[Glossary:Topic|topics]] scheduled for a given day
#'''Burden'''
#*'''Burden''' - current [[Glossary:Burden|''Burden'' parameter]] record. [[Glossary:Burden|Burden]] estimates the number of [[Glossary:Element|elements]] scheduled for [[Glossary:Repetition|repetition]] daily on the assumption of no delays in the learning process. [[Glossary:Burden|Burden]] is the sum of [[Glossary:Interval|interval]] reciprocals. '''[[Glossary:Burden|Burden]] = [[Glossary:Item|Item]] [[Glossary:Burden|burden]] + [[Glossary:Topic|Topic]] [[Glossary:Burden|burden]]'''
#*'''Item burden''' - current '''[[Glossary:Burden|Burden]]''' for [[Glossary:Item|items]]
#*'''Topic burden''' - current '''[[Glossary:Burden|Burden]]''' for [[Glossary:Topic|topics]]
#'''Element count'''
#*'''Elements''' - total number of [[Glossary:Element|elements]] in the [[Glossary:Collection|collection]] as recorded on a given day (i.e. [[Glossary:Topic|topics]] + [[Glossary:Concept|concepts]] + [[Glossary:Item|items]] + [[Glossary:Task|tasks]])
#*'''Items''' - number of [[Glossary:Item|items]] in the [[Glossary:Collection|collection]] as recorded on a given day
#*'''Topics and tasks''' - number of [[Glossary:Topic|topics]] and [[Glossary:Task|tasks]] in the [[Glossary:Collection|collection]] as recorded on a given day (see [[#Use.C2.A0:_Element_count.C2.A0:_Topics_and_tasks|exemplary graph below]])
#*'''Tasklist (''name of the currently selected tasklist'')''' - number of [[Glossary:Task|tasks]] on the currently used [[Glossary:Tasklist|tasklist]] (its name is displayed in parenthesis). Use the '''Set''' button in the [[tasklist manager]] to change the currently selected [[Glossary:Tasklist|tasklist]].
#'''Memorized'''
#*'''Memorized''' - number of [[Glossary:Element|elements]] taking part in the learning process (i.e. not [[Glossary:Dismissed_element|dismissed]] and not [[Glossary:Pending_element|pending]])
#*'''Memorized items''' - number of [[Glossary:Item|items]] taking part in the learning process
#*'''Memorized topics''' - number of [[Glossary:Topic|topics]] taking part in the learning process
#'''Newly memorized'''
#*'''Newly memorized''' - new [[Glossary:Element|elements]] memorized on a given day (see [[#Use.C2.A0:_Newly_memorized|exemplary graph below]])
#*'''New items''' - new [[Glossary:Item|items]] memorized on a given day
#*'''New topics''' - new [[Glossary:Topic|topics]] memorized on a given day
#'''Efficiency'''
#*'''Forgetting index''' - [[Glossary:Measured_forgetting_index|measured forgetting index]] (data stored with a degree of smoothing to visualize trends)
#*'''Retention''' - [[Glossary:Retention|retention]] as measured on each day. This graph is smoothed for better readability of long term trends. For a more precise reading, see the next tab: '''Recall'''. The '''Retention''' tab can now hold over a decade of data and is retained for compatibility. '''Recall''' tab contains unprocessed raw data that you can always polish with '''Smooth''' to visualize trends
#*'''Recall''' - proportion of correct answers on each day (as registered by [https://supermemo.guru/wiki/Algorithm_SM-18 Algorithm SM-18]). This is raw unprocessed data that you can export to Excel for your own analysis. Compare with the '''[[Glossary:Retention|Retention]]''' tab. It should also correspond roughly with data displayed on the '''Forgetting index''' tab (Recall is 100% minus the forgetting index)
#*'''R in SM16''' - [[Glossary:Retrievability|retrievability (R)]] estimated by the [https://supermemo.guru/wiki/Algorithm_SM-15 Algorithm SM-15] (R is the predicted value of [[Glossary:Retention|retention]], while Recall is the measured value of [[Glossary:Retention|retention]])
#*'''R in SM18''' - [[Glossary:Retrievability|retrievability (R)]] estimated by [https://supermemo.guru/wiki/Algorithm_SM-18 Algorithm SM-18]
#*'''SM16/SM18 Diff''' - percentage point difference between recall estimation deviations by SuperMemo algorithms <code>(abs(R16-Recall)-abs(R18-Recall))</code>. The higher the difference, the better the performance of the new [https://supermemo.guru/wiki/Algorithm_SM-18 Algorithm SM-18]. It is the ability of the algorithm to accurately predict the probability of forgetting that makes its strength in [[Glossary:Spaced_repetition|spaced repetition]]. This tab shows you how close [[What's new in SuperMemo 18?|SuperMemo 18]] is to predicting the recall accurately (as compared with older algorithms)
#*'''SM16 Metric''' - performance metric for [https://supermemo.guru/wiki/Algorithm_SM-15 Algorithm SM-15]. It is a square root of the average of prediction differences squared expressed in percentage points (lower number means better recall prediction accuracy)
#*'''SM18 Metric''' - performance metric for [https://supermemo.guru/wiki/Algorithm_SM-18 Algorithm SM-18]. It is computed in the same way as the analogous parameter on the '''SM16 Metric''' tab. Ideally, numbers on this tab should be lower though, indicating superiority of [https://supermemo.guru/wiki/Algorithm_SM-18 Algorithm SM-18]
#*'''R-Metric''' - the ultimate performance metric for [https://supermemo.guru/wiki/Algorithm_SM-18 Algorithm SM-18]. It compares it with the old [https://supermemo.guru/wiki/Algorithm_SM-15 Algorithm SM-15]. It is the performance metric difference between the two metrics from previous tabs: '''SM16 Metric''' minus '''SM18 Metric'''. The higher the number, the greater the superiority of Algorithm SM-18 for a given [[Glossary:Collection|collection]]. On days with a negative metric, the old algorithm would perform better than Algorithm SM-18. As forgetting is random, the actual metric outcome for a day is a lottery. Only a larger number of repetitions per day provide for consistent outcomes. For an exemplary graph based on 8800 data points, see: [[#Use_:_Efficiency_:_R-Metric|example below]].
#*'''Topic load''' - proportion of [[Glossary:Topic|topics]] scheduled for [[Glossary:Review|review]] on a given day among all [[Glossary:Element|elements]] scheduled. ''[[Glossary:Topic|Topic]] load = [[Glossary:Outstanding_element|Outstanding]] [[Glossary:Topic|topics]] / [[Glossary:Outstanding_element|outstanding elements]]''. [[Glossary:Topic|Topic]] load is expressed in percent for any given day. [[Glossary:Topic|Topic]] load should not be confused with '''proportion of [[Glossary:Topic|topics]]''' in [[Priority_queue#Sorting_repetitions|sorting criteria]]. Proportion of [[Glossary:Topic|topics]] is determined independent of the [[Glossary:Topic|topic]] load if possible (e.g. for zero [[Glossary:Topic|topic]] load, you cannot achieve proportion greater than zero, etc.). If you do not use [[Glossary:Auto-sort|auto-sorting]], [[Glossary:Topic|topic]] load tells you what proportion of your repetitions is made of [[Glossary:Topic|topics]].
#'''Overload''' - the number of executed repetitions divided by the number of outstanding repetitions (i.e. the percent of the actually executed repetitions)
#*'''Elements''' - the percent of all outstanding repetitions executed
#*'''Items''' - the percent of [[Glossary:Outstanding_element|outstanding]] [[Glossary:Item|item]] repetitions executed
#*'''Topics''' - the percent of [[Glossary:Outstanding_element|outstanding]] [[Glossary:Topic|topic]] reviews done
#'''Priority protection''' - your processing capacity for high priority material on a given day. If your graph oscillates around [[Glossary:Priority|priority]] of 3%, you will know that only top 3% of your learning material is guaranteed a timely repetition. You can increase that number by doing more work, reducing inflow of new material, deprioritizing less important elements, or reducing the randomization degree for the [[Glossary#Outstanding_queue|outstanding queue]] [[Priority_queue#Sorting_repetitions|sorting criteria]]. Read about this important parameter in '''[[Priority_queue#Prioritization_rulebook|Prioritization Rulebook]]'''
#*'''Items''' - the highest priority [[Glossary:Item|item]] (with the lowest %) that was missed in repetitions. (The actual % in the [[Glossary:Outstanding_queue|outstanding queue]] can be much higher (because you mostly review high [[Glossary:Priority|priorities]]))
#*'''Topics''' - the highest priority [[Glossary:Topic|topic]] (with the lowest %) that was missed in repetitions. (see: [[#Use.C2.A0:_Priority_protection.C2.A0:_Topics|exemplary graph below]])


'''Repetitions''' - distribution of the number of repetitions of items or topics in a given collection; only [[Glossary#memorized_element|memorized elements]] are considered in the distribution, i.e. there is no zero-repetitions category.
=== Use context menu ===


=== Lapses ===
* '''Next''' (''Ctrl+Tab'') - switch to the next major Use statistic tab
* '''Previous''' (''Shift+Ctrl+Tab'') - switch back to the previous major Use statistic tab
* '''Next secondary''' (''Tab'') - switch to the next tab in the secondary row
* '''Previous secondary''' (''Shift+Tab'') - switch to the previous tab in the secondary row
* '''Cut off left''' (''Ctrl+Right arrow'') - cut off the left part of the graph to zoom in on the rest of it
* '''Cut off right''' (''Ctrl+Left arrow'') - cut off the right part of the graph to zoom in on the rest of it
* '''Cut off top''' (''Ctrl+Down arrow'') - cut off the top part of the graph to zoom in on the rest of it
* '''Export''' - export the currently displayed statistic to a CSV file (the file can be found in the [[Glossary:Collection|collection]]'s [STATS] subfolder
* '''Import''' - import data for the currently displayed statistic
* '''Close''' (''Esc'') - close the '''Analysis''' window


'''Lapses''' - distribution of the number of times particular items in the collection have been forgotten; only [[Glossary#memorized_element|memorized elements]] are considered. You should never have more than 10% of items with more than 3-4 lapses. If this is not the case, you should re-examine the way in which you formulate items (see [http://www.supermemo.com/articles/20rules.htm 20 Rules of Formulating Knowledge]). As topics are never "forgotten" at repetition time, there are no lapses listed in '''Lapses''' if you select ''Topics only'' in the element type combo box.
=== Examples ===
[[Image:Lapses distribution.jpg|center|800px|SuperMemo: Tools : Statistics : Analysis : Distributions : Lapses shows you how memory lapses are distributed in your collection]]


== Forgetting <span id="Curves">Curves</span>  ==
==== Use : Efficiency : R-Metric ====


400 forgetting curves are independently plotted in order to compute the RF matrix in [[SuperMemo Algorithm|Algorithm SM-15]]. These correspond to 20 repetition number categories and 20 [[Glossary#A-factor|A-Factor]] categories. For data representation convenience, the columns of the RF matrix for the first repetition are indexed by the number of [[Glossary#lapse|memory lapses]] rather than by the A-Factor. By choosing a proper combination of tabs at the bottom of the graph, you can select a forgetting curve of interest. Horizontal axis represents time expressed as: (1) U-Factor, i.e. the ratio of subsequent inter-repetition intervals, or (2) days (only for the first repetition). Vertical axis represents [[Glossary#retention|knowledge retention]] in percent.  
<div class="center">
[[Image:Recall metric.jpg|800px|SuperMemo: Algorithm SM-18 performance metric]]
{{:Image:Recall metric.jpg}}
</div>


[[Image:Forgetting curves.jpg|center|800px|SuperMemo: Tools : Statistics : Analysis : Forgetting Curves graphs for 20 repetition number categories multiplied by 20 A-Factor categories]]
==== Use : Element count : Topics and tasks ====
<blockquote><span style="padding: 3px; color: #fff; background: blue;">Blue circles</span> in the picture represent data collected during repetitions. The larger the circle, the greater the number of repetitions recorded. The <span style="padding: 3px; color: #fff; background: red;">red curve</span> corresponds with the best-fit forgetting curve obtained by exponential regression. For ill-structured material the forgetting curve is crooked, i.e. not exactly exponential. The horizontal <span style="padding: 3px; color: #fff; background: green;">green line</span> corresponds with the [[Forgetting index#requested_forgetting_index|requested forgetting index]], while the vertical <span style="padding: 3px; color: #fff; background: green;">green line</span> shows the moment in time in which the approximated forgetting curve intersects with the requested forgetting index line. This moment in time determines the value of the relevant [[Glossary#R-Factor|R-Factor]], and indirectly, the value of the optimum interval. For the first repetition, R-Factor corresponds with the first optimum interval. The values of O-Factor and R-Factor are displayed at the top of the graph. They are followed by the number of repetition cases used to plot the graph. At the beginning of the learning process, there is no repetition history and no repetition data to compute R-Factors. It will take some time before your first forgetting curves are plotted. For that reason, the initial value of the RF matrix is taken from the model of a less-than-average student. The model of average student is not used because the convergence from poorer student parameters upwards is faster than the convergence in the opposite direction. The '''Deviation''' parameter displayed at the top tells you how well the negatively exponential curve fits the data. The lesser the deviation, the better the fit. The deviation is computed as a square root of the average of squared differences (as used in the method of least squares).'' </blockquote>
[[Image:Cumulative forgetting curve.jpg|center|800px|SuperMemo: In SuperMemo 16, forgetting curves can be normalized over A-Factors at different repetition categories. As a result, you can display (1) your cumulative forgetting curve (blue dots) and (2) its negative exponential approximation used by SuperMemo (red line).]]
<blockquote>''In SuperMemo 16, forgetting curves can be normalized over A-Factors ('''All AFs''') at different repetition categories ('''All Reps''')(as in the picture above). As a result, you can display:
#''your cumulative forgetting curve (<span style="padding: 3px; color: white; background-color: blue;">blue dots</span>)''
#''its negative exponential approximation used by SuperMemo (<span style="padding: 3px; color: white; background-color: red;">red line</span>).''
'''''Time (normalized)''' stands at 100% for maximum U-factor for each entry of the RF matrix. '''Decay''' stands for the decay constant of the forgetting curve. '''Cases''' shows the number of repetition cases used in plotting the graph. '''Repetitions''' stands for the number of repetitions executed when determining R-Factors. '''Deviation''' is a square root of the average of squared differences between the data and the approximated forgetting curve.''</blockquote>


== 3-D Curves  ==
<div class="center">
[[Image:Topics and tasks count.jpg|800px|SuperMemo: Toolkit : Statistics : Analysis : Use : Element count : Topics and tasks shows you how many topics and tasks there were in your collection on individual days]]
{{:Image:Topics and tasks count.jpg}}
</div>


These graphs show the forgetting curves in 3 dimensions. There are 20 graphs available with A-Factor tabs at the bottom. Each graph corresponds to a different A-Factor (from 1.2 to 6.9). The X axis represents time (as in '''<span id="Forgetting_curves">Forgetting curves</span>'''). The Y axis represents [[Glossary#repetition category|repetition category]] (roughly corresponding to the repetition number). The vertical Z axis represents retention in percent (i.e. how much is remembered at a given moment in time). For better visibility, you can rotate 3-D forgetting curves with '''Rotate''', or smooth them with '''Smooth'''. '''Rotate '''swings&nbsp;the graph&nbsp;on its Z axis by 90 degrees clockwise&nbsp;(as&nbsp;viewed from above).&nbsp;This means&nbsp;that X and Y axes get swapped.&nbsp;Note that smoothing proceeds along the Y axis. This makes it possible to align incompletely plotted forgetting curves from different repetition categories. However, once you click '''Rotate''', Y axis smoothing will produce a different result (e.g. after the first '''Rotate''', smoothing will reduce kinks along the repetition categories). For the best illustration of the forgetting process, smooth the graph before you rotate it. <!-- Uhhh.... better explain X- and Y-axes; which one goes into the screen? And what is "repetition category"? -->
==== Use : Newly memorized ====
 
[[Image:AFactor_3D_Curve.jpg|center|800px|SuperMemo: Exemplary 3-D graph of forgetting curves for A-Factor=3.6]]


<div class="center">
<div class="center">
''Exemplary 3-D graph of forgetting curves for A-Factor=3.6. The graph has been smoothed along the Y-axis (by averaging results corresponding to different repetition categories), and rotated once for better visibility (the slope illustrates forgetting in time that is roughly similar for various repetition categories except those with high repetition number that may reflect scarcity of data for long intervals).''
[[Image:Newly memorized.jpg|800px|SuperMemo: Toolkit : Statistics : Analysis : Use : Newly memorized shows you how many elements you memorized on individual days]]
{{:Image:Newly memorized.jpg}}
</div>
</div>


== Graphs  ==
==== Use : Efficiency : Forgetting index ====  


=== Forgetting Index vs. Priority ===
<div class="center">
[[Image:Daily measured forgetting index.jpg|800px|SuperMemo: Toolkit : Statistics : Analysis : Use : Efficiency : Forgetting index shows you the changes to the measured forgetting index on individual days]]
{{:Image:Daily measured forgetting index.jpg}}
</div>


'''Forgetting Index vs. Priority''' - this graph shows you how forgetting depends on the [[Priority queue|priority]] of the learning material. If you always go through your outstanding material in time and if you stick to the [http://supermemo.com/articles/20rules.htm 20 rules of knowledge formulation], your [[Glossary#measured_forgetting_index|measured forgetting index]] should be the same as the [[Glossary#requested_forgetting_index|requested forgetting index]]. This graph will then show a flat straight line passing horizontally at the level of the forgetting index. However, once you start overloading the learning process and using '''[[Postpone]]''' or '''[[Mercy]]''', your measured forgetting index will increase. If you always sort your repetitions by priority (e.g. by leaving '''[[Learn menu|Learn]]&nbsp;: Sorting&nbsp;: Auto-sort repetitions''' checked), you should still ensure that your measured forgetting index equals the requested forgetting index for high priority elements (i.e. on the left side of the graph). Your measured forgetting index for low priority material will be higher, and the greater the overload the greater the difference. If you change your repetition sorting criteria by increasing randomization of repetitions, this graph will be flatter and the measured forgetting index for high priority material will be higher.
==== Use : Priority protection : Items ====


=== First interval ===
<div class="center">
[[Image:Item priority protection.jpg|800px|SuperMemo: Toolkit : Statistics : Analysis : Use : Priority protection : Items shows your actual processing capacity for high-priority items on individual days]]
{{:Image:Item priority protection.jpg}}
</div>


'''First interval''' - the length of the first interval after the first repetition depends on the number of times a given item has been forgotten. Note that the first repetition here means the first repetition after forgetting, ''not'' the first repetition ever. In other words, a twice repeated item will have the repetition number equal to one after it has been forgotten; the repetition number will not equal three. The first interval graph shows exponential regression curve that approximates the length of the first interval for different numbers of [[Glossary#lapse|memory lapses]] (including the zero-lapses category that corresponds with newly memorized items). In the graph below, blue circles correspond to data collected in the learning process (the greater the circle, the more repetitions have been recorded).
==== Use : Priority protection : Topics ====


[[Image:First interval.jpg|center|800px|SuperMemo: Tools : Statistics : Analysis : Graphs : First Interval shows you exponential regression curve that approximates the length of the first interval for different numbers of memory lapses]]
<div class="center">
<div class="center">
''In the graph above, which includes data from over 130,000 repetitions, newly memorized items are optimally repeated after seven days. However, items that have been forgotten 10 times (which is rare in SuperMemo) will require an interval of two days. (Due to logarithmic scaling, the size of the circle is not linearly proportional to the data sample; the number of repetition cases for lapses=0 is by far larger than for lapses=10, as can be seen in '''Distributions&nbsp;: Lapses'''.)''
[[Image:Topic priority protection.jpg|800px|SuperMemo: Toolkit : Statistics : Analysis : Use : Priority protection : Topics shows your actual processing capacity for high-priority topics on individual days]]
{{:Image:Topic priority protection.jpg}}
</div>
</div>


=== D-Factor vs. A-Factor ===
== Forgetting ==
 
400 [[Glossary:Forgetting_curve|forgetting curves]] are independently plotted in order to compute the SInc matrix in [https://supermemo.guru/wiki/Algorithm_SM-18 Algorithm SM-18]. These correspond to 20 stability categories and 20 difficulty categories. By choosing a proper combination of tabs at the bottom of the graph, you can select a [[Glossary:Forgetting_curve|forgetting curve]] of interest. Horizontal axis represents time expressed as: (1) retrievability, or (2) days (only for the first [[Glossary:Repetition|repetition]]). Vertical axis represents [[Glossary:Retention|knowledge retention]] in percent.
 
[[Image:First forgetting curve.jpg|center|600px|SuperMemo: The first forgetting curve (first repetition for items with no lapses)]]
{{:Image:First forgetting curve.jpg}}
 
== Forgetting (UF)  ==


'''D-Factor vs. A-Factor''' - DF-AF graph shows decay constants of the power approximation of R-Factors along columns of the RF matrix. You need to understand [[SuperMemo Algorithm]] to understand this graph. The horizontal axis represents [[Glossary#A-factor|A-Factor]], while the vertical axis represents D-Factor (i.e. Decay Factor). D-Factor is a decay constant of the power approximation of curves that can be inspected in the '''Approximations''' tab of the '''Analysis''' dialog box.
This tab displays forgetting data as used in older SuperMemos. It is only used to record and display comparisons between the new algorithm and the older algorithms.


=== First Grade vs. A-Factor ===
400 [[Glossary:Forgetting_curve|forgetting curves]] are independently plotted in order to compute the RF matrix in [https://supermemo.guru/wiki/Algorithm_SM-15 Algorithm SM-15]. These correspond to 20 repetition number categories and 20 [[Glossary:A-Factor|A-Factor]] categories. For data representation convenience, the columns of the RF matrix for the first repetition are indexed by the number of [[Glossary:Lapse|memory lapses]] rather than by the [[Glossary:A-Factor|A-Factor]]. By choosing a proper combination of tabs at the bottom of the graph, you can select a [[Glossary:Forgetting_curve|forgetting curve]] of interest. Horizontal axis represents time expressed as: (1) [[Glossary:U-Factor|U-Factor]], i.e. the ratio of subsequent inter-repetition [[Glossary:Interval|intervals]], or (2) days (only for the first [[Glossary:Repetition|repetition]]). Vertical axis represents [[Glossary:Retention|knowledge retention]] in percent.


'''First Grade vs. A-Factor''' - G-AF graph correlates the first grade obtained by an item with the ultimate estimation of its [[Glossary#A-factor|A-Factor]] value. At each repetition, the current element's old A-Factor estimation is removed from the graph and the new estimation is added. This graph is used by [[SuperMemo Algorithm]] to quickly estimate the first value of A-Factor at the moment when all we know about an element is the first grade it has scored in its first repetition.
[[Image:Forgetting curves.jpg|600px|center|SuperMemo: Toolkit : Statistics : Analysis : Forgetting (UF) graphs for 20 repetition number categories multiplied by 20 A-Factor categories]]  
{{:Image:Forgetting curves.jpg}}


=== Grade vs. Forgetting Index ===
[[Image:Cumulative forgetting curve.jpg|center|800px|SuperMemo: In SuperMemo 18, forgetting curves can be normalized over A-Factors at different repetition categories. As a result, you can display (1) your cumulative forgetting curve (blue dots), (2) its negative exponential approximation used by SuperMemo (red line) and (3) individual data points from all forgetting curves collected by SuperMemo (yellow circles).]]
{{:Image:Cumulative forgetting curve.jpg}}


'''<span id="G-FI_graph">Grade vs. Forgetting Index</span>''' - FI-G graph correlates the expected forgetting index with the grade scored at repetitions. You need to understand [[SuperMemo Algorithm]] to understand this graph. You can imagine that the forgetting curve graph might use the average grade instead of the [[Glossary#retention|retention]] on its vertical axis. If you correlated this grade with the [[Forgetting index|forgetting index]], you would arrive at the FI-G graph. This graph is used to compute an estimated forgetting index that is in turn used to normalize grades (for delayed or advanced repetitions) and estimate the new value of item's A-Factor. The grade is computed using the formula: ''Grade=Exp(A*FI+B)'', where A and B are parameters of an exponential regression run over raw data collected during repetitions.
== Stabilization ==


[[Image:Grade vs Forgetting index.jpg|center|800px|SuperMemo: Tools : Statistics : Analysis : Graphs : Grade vs. Forgetting index graph]]
The tab marked '''Stabiliz.''' in '''Analysis''' can be used to inspect [[Glossary:stabilization curve|stabilization curve]]s in SuperMemo.


== 3-D Graphs  ==
[[File:Stabilization curve.jpg|800px|center|Inspecting stabilization curves for various levels of stability and difficulty]]
{{:File:Stabilization curve.jpg}}


3-D graphs that visually represent RF, OF, and Cases matrices (see: [[#Matrices|'''Matrices''']] below)
== Graphs  ==


== Approximations  ==
=== Forgetting Index vs. Priority ===


20 power approximation curves that show the decline of R-Factors along columns of the RF matrix. You need to understand [[SuperMemo Algorithm]] to understand these graphs. For each [[Glossary#A-factor|A-Factor]], with increasing values of the repetition number, the value of R-Factor decreases (at least theoretically it should decrease). Power regression is used to illustrate the degree of this decline that is best reflected by the decay constant called here D-Factor. By choosing the A-Factor tab at the bottom of the graph, you can view a corresponding R-Factor approximation curve. The horizontal axis represents the repetition number category, while the vertical axis represents R-Factor. The value of D-Factor is shown at the top of the graph. The blue polyline shows R-Factors as derived from repetition data. The red curve shows the fixed-point power approximation of R-Factor. The green curve shows the fixed-point power approximation of R-Factor derived from the OF matrix; this is equivalent to substituting the D-Factor obtained by fixed-point power approximation of R-Factors for the D-Factor obtained from DF-AF linear regression. A fixed-point approach is used for both approximations since for the repetition number equal to two, R-Factor equals A-Factor.
'''Forgetting Index vs. Priority''' - this graph shows you how forgetting depends on the [[Glossary:Priority|priority]] of the learning material. If you always go through your [[Glossary:Outstanding_element|outstanding material]] in time and if you stick to the [http://super-memory.com/articles/20rules.htm 20 rules of knowledge formulation], your [[Glossary:Measured_forgetting_index|measured forgetting index]] should be the same as the [[Glossary:Requested_forgetting_index|requested forgetting index]]. This graph will then show a flat straight line passing horizontally at the level of the [[Glossary:Forgetting_index|forgetting index]]. However, once you start overloading the learning process and using '''[[Postpone]]''' or '''[[Mercy]]''', your [[Glossary:Measured_forgetting_index|measured forgetting index]] will increase. If you always sort your repetitions by [[Glossary:Priority|priority]] (e.g. by leaving '''[[Learn_menu|Learn]] : [[Learn_menu#Sorting|Sorting]] : Auto-sort repetitions''' checked), you should still ensure that your [[Glossary:Measured_forgetting_index|measured forgetting index]] equals the [[Glossary:Requested_forgetting_index|requested forgetting index]] for high priority [[Glossary:Element|elements]] (i.e. on the left side of the graph). Your [[Glossary:Measured_forgetting_index|measured forgetting index]] for low priority material will be higher, and the greater the [[Glossary:Overload|overload]] the greater the difference. If you change your [[Priority_queue#Sorting_repetitions|repetition sorting criteria]] by increasing randomization of repetitions, this graph will be flatter and the [[Glossary:Measured_forgetting_index|measured forgetting index]] for high priority material will be higher.


== Matrices  ==
=== First interval ===


You need to understand [[SuperMemo Algorithm]] to understand these tables.  
'''First interval''' - the length of the first [[Glossary:Interval|interval]] after the first [[Glossary:Repetition|repetition]] depends on the number of times a given [[Glossary:Item|item]] has been forgotten. Note that the first [[Glossary:Repetition|repetition]] here means the first [[Glossary:Repetition|repetition]] after forgetting, ''not'' the first repetition ever. In other words, a twice repeated [[Glossary:Item|item]] will have the repetition number equal to one after it has been forgotten; the repetition number will not equal three. The first [[Glossary:Interval|interval]] graph shows exponential regression curve that approximates the length of the first [[Glossary:Interval|interval]] for different numbers of [[Glossary:Lapse|memory lapses]] (including the zero-lapses category that corresponds with newly [[Glossary:memorized_element|memorized]] [[Glossary:Item|items]]). In the graph below, blue circles correspond to data collected in the learning process (the greater the circle, the more repetitions have been recorded).


* '''OF matrix''' - matrix of optimal factors indexed by the repetition number and [[Glossary#A-factor|A-Factor]] (except for the first repetition, A-Factor is replaced with memory lapses)
<div class="center">
* '''RF matrix''' - matrix of retention factors
[[Image:First interval.jpg|800px|SuperMemo: Toolkit : Statistics : Analysis : Graphs : First Interval shows you exponential regression curve that approximates the length of the first interval for different numbers of memory lapses]]
* '''Cases''' - matrix of repetition cases used to compute the corresponding entries of the RF matrix (double click an entry to view the relevant forgetting curve). This matrix can be edited manually (e.g. if you want to change the weight of some measurements during repetitions)
{{:Image:First interval.jpg}}
* '''Intervals''' - matrix of optimum intervals derived from the OF matrix
</div>
* '''D-Factors''' - vector of D-Factor values for different [[Glossary#A-factor|A-Factor]] values (also repetition cases used in computing particular D-Factors)


== Use  ==
=== D-Factor vs. A-Factor ===


Graphs of changes in the statistics of the learning process over time. These graphs record daily changes to statistics such as the number of memorized items, measured forgetting index or the number of outstanding items for a given day. The first non-zero value of the day is recorded. If that value increases afterwards, that increase is ignored (except a few cases such as the statistic of newly memorized items, etc.). If you have a break in learning, the last recorded non-zero value will propagate through the period in which you did not use SuperMemo (except a few cases such as the statistics of repetitions made, etc.).  
'''D-Factor vs. A-Factor''' - DF-AF graph shows decay constants of the power approximation of [[Glossary:R-Factor|R-Factors]] along columns of the [[Glossary:RF_matrix|RF matrix]]. You need to understand [https://supermemo.guru/wiki/Algorithm_SM-15 SuperMemo Algorithm SM-15] to understand this graph. The horizontal axis represents [[Glossary:A-Factor|A-Factor]], while the vertical axis represents [[Glossary:D-Factor|D-Factor]] (i.e. Decay Factor). [[Glossary:D-Factor|D-Factor]] is a decay constant of the power approximation of curves that can be inspected in the '''[[#Approximations|Approximations]]''' tab of the '''Analysis''' dialog box.


To zoom in on a portion of the graph (as in pictures below), sweep the portions of the graphs that are to be removed with the mouse, i.e. point at the day of focus, move the mouse left and release the mouse button. This cuts away part of the graph in the direction that the mouse was moved. You can also use ''Ctrl+Left'' and ''Ctrl+Right'' buttons to zoom in (see [[#Use context menu | below]])
=== First Grade vs. A-Factor ===


=== Use statistics ===
'''First Grade vs. A-Factor''' - G-AF graph correlates the first [[Glossary:Grade|grade]] obtained by an [[Glossary:Item|item]] with the ultimate estimation of its [[Glossary:A-Factor|A-Factor]] value. At each [[Glossary:Repetition|repetition]], the current [[Glossary:Element|element]]'s old [[Glossary:A-Factor|A-Factor]] estimation is removed from the graph and the new estimation is added. This graph is used by [https://supermemo.guru/wiki/Algorithm_SM-15 SuperMemo Algorithm SM-15] to quickly estimate the first value of [[Glossary:A-Factor|A-Factor]] at the moment when all we know about an [[Glossary:Element|element]] is the first [[Glossary:Grade|grade]] it has scored in its first [[Glossary:Repetition|repetition]].


Here are the statistics recorded daily by SuperMemo and displayed on the '''Use''' tab:
=== Grade vs. Forgetting Index ===


#'''Work done'''
'''<span id="G-FI_graph">Grade vs. Forgetting Index</span>''' - FI-G graph correlates the [https://supermemo.guru/wiki/Forgetting_index_in_SuperMemo#expected_forgetting_index expected forgetting index] with the [[Glossary:Grade|grade]] scored at repetitions. You need to understand [https://supermemo.guru/wiki/Algorithm_SM-15 SuperMemo Algorithm SM-15] to understand this graph. You can imagine that the [[Glossary:Forgetting_curve|forgetting curve]] graph might use the average [[Glossary:Grade|grade]] instead of the [[Glossary:Retention|retention]] on its vertical axis. If you correlated this grade with the [[Glossary:Forgetting_index|forgetting index]], you would arrive at the FI-G graph. This graph is used to compute an [https://supermemo.guru/wiki/Forgetting_index_in_SuperMemo#estimated_forgetting_index estimated forgetting index] that is in turn used to normalize grades (for delayed or advanced repetitions) and estimate the new value of [[Glossary:Item|item]]'s [[Glossary:A-Factor|A-Factor]]. The grade is computed using the formula: ''Grade=Exp(A*FI+B)'', where A and B are parameters of an exponential regression run over raw data collected during repetitions.
#*'''Use time''' - your daily total time in which you actively used a given collection (as recorded in '''[[Tools menu|Tools]] : Sleep Chart''')
#*'''Recall time''' - your daily total time which you spent on trying to recall answers to questions
#*'''Repetitions''' - your daily recorded number of item and topic reviews
#*'''Item repetitions''' - your daily record of item repetitions
#*'''Topic repetitions''' - your daily record of topic reviews
#'''Outstanding''' - number of outstanding elements scheduled for a given day
#*'''Outstanding''' - number of all outstanding elements scheduled for a given day
#*'''Outstanding items''' - number of outstanding items scheduled for a given day
#*'''Outstanding topics''' - number of outstanding topics scheduled for a given day
#'''Burden'''
#*'''Burden''' - current [[Glossary#burden|''Burden'' parameter]] record. Burden estimates the number of elements scheduled for repetition daily on the assumption of no delays in the learning process. Burden is the sum of interval reciprocals. '''Burden = Item burden + Topic burden'''
#*'''Item burden''' - current '''Burden''' for items
#*'''Topic burden''' - current '''Burden''' for topics
#'''Element count'''
#*'''Elements''' - total number of elements in the collection as recorded on a given day (i.e. topics + items + tasks)
#*'''Items''' - number of items in the collection as recorded on a given day
#*'''Topics and tasks''' - number of topics and tasks in the collection as recorded on a given day (see an exemplary graph below)
#*'''Tasklist (''name of the currently selected tasklist'')''' - number of [[Glossary#task|tasks]] on the currently used [[Glossary#tasklist|tasklist]] (its name is displayed in parenthesis). Use the '''Tasklist''' combo box in the [[Tools toolbar|'''Tools''' toolbar]] to change the currently selected tasklist.
#'''Memorized'''
#*'''Memorized''' - number of elements taking part in the learning process (i.e. not dismissed and not pending)
#*'''Memorized items''' - number of items taking part in the learning process
#*'''Memorized topics''' - number of topics taking part in the learning process
#'''Newly memorized'''
#*'''Newly memorized''' - new elements memorized on a given day (see an exemplary graph below)
#*'''New items''' - new items memorized on a given day
#*'''New topics''' - new topics memorized on a given day
#'''Efficiency'''
#*'''Forgetting index''' - [[Forgetting index|measured forgetting index]]  
#*'''Retention''' - retention as measured on each day
#*'''Topic load''' - proportion of topics scheduled for review on a given day among all elements scheduled. ''Topic load = Outstanding topics / Outstanding elements''. Topic load is expressed in percent for any given day.Topic should not be confused with '''proportion of topics''' in sorting criteria. Proportion of topics is determined independent on the topic load if possible (e.g. for zero topic load, you cannot achieve proportion greater than zero, etc.). If you do not use auto-sorting, topic load tells you what proportion of your repetitions is made of topics.
#'''Overload''' - the number of executed repetitions divided by the number of outstanding repetitions (i.e. the percent of the actually executed repetitions)  
#*'''Elements''' - the percent of all outstanding repetitions executed
#*'''Items''' - the percent of outstanding item repetitions executed
#*'''Topics''' - the percent of outstanding topic reviews done
#'''Priorities missed''' - your processing capacity for high priority material on a given day. If your graph oscillates around Priority of 3%, you will know that only top 3% of your learning material is guaranteed a timely repetition. You can increase that number by doing more work, reducing inflow of new material, deprioritizing less important elements, or reducing the randomization degree for the outstanding queue sorting criteria. Read about this important parameter in '''Prioritization Rulebook''' in the [[priority queue]] article
#*'''Items''' - the highest priority item (with the lowest&nbsp;%) that was missed in repetitions. (The actual&nbsp;% in the outstanding queue can be much higher (because you mostly review high priorities))
#*'''Topics''' - the highest priority topic (with the lowest&nbsp;%) that was missed in repetitions. (see an exemplary graph below)


=== Use context menu ===
[[Image:Grade vs Forgetting index.jpg|center|800px|SuperMemo: Toolkit : Statistics : Analysis : Graphs : Grade vs. Forgetting index graph]]


* '''Next''' (''Ctrl+Tab'') - switch to the next major Use statistic tab
== Distributions ==
* '''Previous''' (''Shift+Ctrl+Tab'') - switch back to the previous major Use statistic tab
* '''First''' (''Home'') - switch to the first major Use statistic tab ('''Work done''' in SuperMemo 15)


* '''Cut off left''' (''Ctrl+Right arrow'') - cut off the left part of the graph to zoom in on the rest of it
=== Intervals ===
* '''Cut off right''' (''Ctrl+Left arrow'') - cut off the right part of the graph to zoom in on the rest of it
* '''Cut off top''' (''Ctrl+Down arrow'') - cut off the top part of the graph to zoom in on the rest of it


* '''Close''' (''Esc'') - close the '''Analysis''' window
'''Intervals''' - distribution of inter-repetition [[Glossary:Interval|intervals]] in a given [[Glossary:Collection|collection]]. By changing the setting of the combo box at the bottom of the dialog, you can see [[Glossary:Interval|interval]] distributions for all [[Glossary:Element|elements]] in the [[Glossary:Collection|collection]], for only [[Glossary:Item|items]], and for only [[Glossary:Topic|topics]].


=== Examples ===
=== A-Factors ===


==== Use : Element count : Topics and tasks ====
'''A-Factors''' - distribution of [[Glossary#A-factor|A-Factors]] in a given [[Glossary:Collection|collection]]. The distribution itself is not used in [https://supermemo.guru/wiki/Algorithm_SM-15 Algorithm SM-15], and merely results from it. Note that [[Glossary:Topic|topics]] and [[Glossary:Item|items]] use different [[Glossary:A-Factor|A-Factor]] categories. Choose ''Items only'' or ''Topics only'' at the bottom to see distributions for items or topics respectively. If you choose ''All elements'' you will see a less meaningful conglomerate distribution indexed by [[Glossary:A-Factor|A-Factor]] category number.
[[Image:AFactors distribution.jpg|center|800px|SuperMemo: Toolkit : Statistics : Analysis : Distributions : A-Factors shows you how A-Factors are distributed in your collection]]


[[Image:Topics and tasks count.jpg|center|800px|SuperMemo: Tools : Statistics : Analysis : Use : Element count : Topics and tasks shows you how many topics and tasks there were in your collection on individual days]]
=== Difficulties ===


<blockquote>
'''Difficulties''' - distribution of item difficulties as determined by [https://supermemo.guru/wiki/Algorithm_SM-18 Algorithm SM-18]. See also: [[Glossary:Difficulty|difficulties]]


''Exemplary graph showing the total number of topics stored in the collection. The data displayed at the bottom depends on the position of the mouse. In this example, it indicates that on Mar 6, 2009, there were 251,228 topics in the incremental reading process.''
=== Repetitions ===


</blockquote>
'''Repetitions''' - distribution of the number of [[Glossary:Repetition|repetitions]] of [[Glossary:Item|items]] or [[Glossary:Topic|topics]] in a given [[Glossary:Collection|collection]]; only [[Glossary:Memorized_element|memorized elements]] are considered in the distribution, i.e. there is no zero-repetitions category.


==== Use : Newly memorized ====
=== Lapses ===


[[Image:Newly memorized.jpg|center|800px|SuperMemo: Tools : Statistics : Analysis : Use : Newly memorized shows you how many elements you memorized on individual days]]
'''Lapses''' - distribution of the number of times particular [[Glossary:Item|items]] in the [[Glossary:Collection|collection]] have been forgotten; only [[Glossary:Memorized_element|memorized elements]] are considered. You should never have more than 10% of [[Glossary:Item|items]] with more than 3-4 lapses. If this is not the case, you should re-examine the way in which you formulate [[Glossary:Item|items]] (see [http://www.super-memory.com/articles/20rules.htm 20 Rules of Formulating Knowledge]). As [[Glossary:Topic|topics]] are never "forgotten" at repetition time, there are no [[Glossary:Lapse|lapses]] listed in '''Lapses''' if you select ''Topics only'' in the element type combo box.
[[Image:Lapses distribution.jpg|center|800px|SuperMemo: Toolkit : Statistics : Analysis : Distributions : Lapses shows you how memory lapses are distributed in your collection]]


<blockquote>
== 3-D Curves  ==


''Exemplary graph showing the number of elements memorized on individual days. By sweeping away the left and right part of the graph, it is possible to zoom in on a small section of the learning process. The period shown on the graph spans from Jul 1, 2008 to Mar 19, 2009.''
These graphs show the [[Glossary:Forgetting_curve|forgetting curves]] in 3 dimensions. There are 20 graphs available with [[Glossary:A-Factor|A-Factor]] tabs at the bottom. Each graph corresponds to a different [[Glossary:A-Factor|A-Factor]] (from 1.2 to 6.9). The X axis represents time (as in '''[[#Forgetting curves|Forgetting curves]]'''). The Y axis represents [[Glossary:Repetition_category|repetition category]] (roughly corresponding to the repetition number). The vertical Z axis represents [[Glossary:Retention|retention]] in percent (i.e. how much is remembered at a given moment in time). For better visibility, you can rotate 3-D [[Glossary:Forgetting_curve|forgetting curves]] with '''Rotate''', or smooth them with '''Smooth'''. '''Rotate '''swings the graph on its Z axis by 90 degrees clockwise (as viewed from above). This means that X and Y axes get swapped. Note that smoothing proceeds along the Y axis. This makes it possible to align incompletely plotted [[Glossary:Forgetting_curve|forgetting curves]] from different [[Glossary:Repetition_category|repetition categories]]. However, once you click '''Rotate''', Y axis smoothing will produce a different result (e.g. after the first '''Rotate''', smoothing will reduce kinks along the [[Glossary:Repetition_category|repetition categories]]). For the best illustration of the forgetting process, smooth the graph before you rotate it. <!-- Uhhh.... better explain X- and Y-axes; which one goes into the screen? And what is "repetition category"? -->


</blockquote>
<div class="center">
[[Image:AFactor_3D_Curve.jpg|800px|SuperMemo: Exemplary 3-D graph of forgetting curves for A-Factor=3.6]]
{{:Image:AFactor_3D_Curve.jpg}}
</div>


==== Use : Efficiency : Forgetting index ====  
== 3-D Graphs  ==


[[Image:Daily measured forgetting index.jpg|center|800px|SuperMemo: Tools : Statistics : Analysis : Use : Efficiency : Forgetting index shows you the changes to the measured forgetting index on individual days]]
3-D graphs that visually represent [[Glossary:R-Factor|RF]], [[Glossary:O-Factor|OF]], and Cases matrices (see: [[#Matrices|'''Matrices''']] below)


<blockquote>
== S. decay ==


''Exemplary graph enabling a more meaningful analysis of the forgetting index: changes to forgetting index in '''Analysis''' use the daily measured forgetting index (previously: less informative cumulative measured forgetting index value was used taken for the entire period since the last use of '''[[Tools menu|Tools]] : Statistics&nbsp;: Reset parameters&nbsp;: Forgetting index record'''). Note that the priority queue may distort the actual retention in your collection as measured values are primarily taken from top-priority material. Thus measured forgetting index should be understood as "forgetting index measured at repetitions", not as "overall measured forgetting index".''
Graphs that show how SInc slope is computed.


</blockquote>
== OF decay ==


==== Use : Priorities missed : Topics ====
20 power approximation curves that show the decline of [[Glossary:R-Factor|R-Factors]] along columns of the [[Glossary:RF_matrix|RF matrix]]. You need to understand [https://supermemo.guru/wiki/Algorithm_SM-15 SuperMemo Algorithm SM-15] to understand these graphs. For each [[Glossary:A-Factor|A-Factor]], with increasing values of the repetition number, the value of [[Glossary:R-Factor|R-Factor]] decreases (at least theoretically it should decrease). Power regression is used to illustrate the degree of this decline that is best reflected by the decay constant called here [[Glossary:D-Factor|D-Factor]]. By choosing the [[Glossary:A-Factor|A-Factor]] tab at the bottom of the graph, you can view a corresponding [[Glossary:R-Factor|R-Factor]] approximation curve. The horizontal axis represents the repetition number category, while the vertical axis represents [[Glossary:R-Factor|R-Factor]]. The value of [[Glossary:D-Factor|D-Factor]] is shown at the top of the graph. The blue polyline shows [[Glossary:R-Factor|R-Factors]] as derived from repetition data. The red curve shows the fixed-point power approximation of [[Glossary:R-Factor|R-Factor]]. The green curve shows the fixed-point power approximation of [[Glossary:R-Factor|R-Factor]] derived from the OF matrix; this is equivalent to substituting the [[Glossary:D-Factor|D-Factor]] obtained by fixed-point power approximation of [[Glossary:R-Factor|R-Factors]] for the [[Glossary:D-Factor|D-Factor]] obtained from DF-AF linear regression. A fixed-point approach is used for both approximations since for the repetition number equal to two, [[Glossary:R-Factor|R-Factor]] equals [[Glossary:A-Factor|A-Factor]].


[[Image:Priorities missed Topics.jpg|center|800px|SuperMemo: Tools : Statistics : Analysis : Use : Priorities missed : Topics shows your actual processing capacity for high-priority topics on individual days]]
== Matrices  ==


<blockquote>
You need to understand [[SuperMemo Algorithm]] to understand these tables.


''Exemplary graph showing the highest priority topic (with the lowest priority&nbsp;%) that was missed in repetitions. The data displayed at the bottom depends on the position of the mouse. In this example, it indicates that on Feb 13, 2009, the worst miss amongst topics was a topic with Priority equal to 3.11% standing at the position 85 of the queue of the most important topics to review.''
* '''Recall''' - recall derived from [[Glossary:Recall matrix|Recall[] matrix]]
* '''Metric''' - recall metric that shows the accuracy of recall predictions for different levels of [[Glossary:Stability|stability]] and [[Glossary:Difficulty|difficulty]]
* '''Cases''' - repetition cases used in computing the stabilization matrix
* '''R90''' - predicted [[Glossary:Retrievability|retrievability]] at recall of 90%
* '''SInc''' - stabilization expressed as [[Glossary:Stability increase matrix|stability increase matrix SInc[]]]
* '''Intervals''' - matrix of [[Glossary:Optimum_interval|optimum intervals]] derived from the OF matrix
* '''R-Factors''' - matrix of [[Glossary:R-Factor|retention factors]]
* '''OF matrix''' - matrix of [[Glossary:O-Factor|optimal factors]] indexed by the repetition number and [[Glossary:A-Factor|A-Factor]] (except for the first [[Glossary:Repetition|repetition]], [[Glossary:A-Factor|A-Factor]] is replaced with [[Glossary:Lapse|memory lapses]])
* '''RF Cases''' - matrix of repetition cases used to compute the corresponding entries of the [[Glossary:RF_matrix|RF matrix]] (double click an entry to view the relevant [[Glossary:Forgetting_curve|forgetting curve]]). This matrix can be edited manually (e.g. if you want to change the weight of some measurements during repetitions)


</blockquote>
[[File:Intervals matrix.jpg|center|600px|SuperMemo: Toolkit : Statistics : Matrices : Intervals displays the matrix of optimum intervals derived from the OF matrix]]


== FAQ ==
== FAQ ==


* [http://wiki.supermemo.org/index.php?title=Weirdness_in_my_collection I worry my graphs look weird and I do something wrong]
* [https://supermemopedia.com/wiki/Weirdness_in_my_collection I worry my graphs look weird and I do something wrong]
* [https://supermemopedia.com/wiki/Random_numbers_showed_up_in_data_used_by_the_algorithm Errors in data in Analysis]


== Further reading  ==
== Further reading  ==
 
*[[Memory graphs (4D)]]
*[http://supermemo.com/help/faq/memory.htm FAQ: Memory and learning]  
*[https://supermemo.guru/wiki/SuperMemo_Algorithm SuperMemo Algorithm]
*[http://supermemo.com/articles/theory.htm SuperMemo Theory]  
*[https://supermemo.guru/wiki/Algorithm_SM-15 SuperMemo Algorithm SM-15]
*[http://supermemo.com/english/algsm11.htm SuperMemo Algorithm]
*[http://super-memory.com/help/faq/memory.htm FAQ: Memory and learning]  
*[http://super-memory.com/articles/theory.htm SuperMemo Theory]

Revision as of 06:30, 25 July 2019

The tabbed dialog available with Toolkit : Statistics : Analysis (Shift+Alt+A) in SuperMemo provides graphs and tables that illustrate the current state of the learning process in the currently opened collection.

In addition to Analysis, SuperMemo provides multiple other analytical tools.

Use

Graphs of changes in the statistics of the learning process over time. These graphs record daily changes to statistics such as the number of memorized items, measured forgetting index or the number of outstanding items for a given day. The first non-zero value of the day is recorded. If that value increases afterwards, that increase is ignored (except a few cases such as the statistic of newly memorized items, etc.). If you have a break in learning, the last recorded non-zero value will propagate through the period in which you did not use SuperMemo (except a few cases such as the statistics of repetitions made, etc.).

To zoom in on a portion of the graph (as in pictures below), sweep the portions of the graphs that are to be removed with the mouse, i.e. point at the day of focus, move the mouse left and release the mouse button. This cuts away part of the graph in the direction that the mouse was moved. You can also use Ctrl+Left and Ctrl+Right buttons to zoom in (see below)

Use statistics

Here are the statistics recorded daily by SuperMemo and displayed on the Use tab:

  1. Work done
    • Use time - your daily total time in which you actively used a given collection. The time is taken from learning blocks record (Toolkit : Sleep Chart in red). Days count from midnight to midnight (i.e. midnight shift is disregarded)
    • Recall time - your daily total time which you spent on trying to recall answers to questions
    • Repetitions - your daily recorded number of item and topic reviews
    • Item repetitions - your daily record of item repetitions
    • Topic reviews - your daily record of topic reviews
  2. Outstanding - number of outstanding elements scheduled for a given day
  3. Burden
  4. Element count
  5. Memorized
    • Memorized - number of elements taking part in the learning process (i.e. not dismissed and not pending)
    • Memorized items - number of items taking part in the learning process
    • Memorized topics - number of topics taking part in the learning process
  6. Newly memorized
  7. Efficiency
    • Forgetting index - measured forgetting index (data stored with a degree of smoothing to visualize trends)
    • Retention - retention as measured on each day. This graph is smoothed for better readability of long term trends. For a more precise reading, see the next tab: Recall. The Retention tab can now hold over a decade of data and is retained for compatibility. Recall tab contains unprocessed raw data that you can always polish with Smooth to visualize trends
    • Recall - proportion of correct answers on each day (as registered by Algorithm SM-18). This is raw unprocessed data that you can export to Excel for your own analysis. Compare with the Retention tab. It should also correspond roughly with data displayed on the Forgetting index tab (Recall is 100% minus the forgetting index)
    • R in SM16 - retrievability (R) estimated by the Algorithm SM-15 (R is the predicted value of retention, while Recall is the measured value of retention)
    • R in SM18 - retrievability (R) estimated by Algorithm SM-18
    • SM16/SM18 Diff - percentage point difference between recall estimation deviations by SuperMemo algorithms (abs(R16-Recall)-abs(R18-Recall)). The higher the difference, the better the performance of the new Algorithm SM-18. It is the ability of the algorithm to accurately predict the probability of forgetting that makes its strength in spaced repetition. This tab shows you how close SuperMemo 18 is to predicting the recall accurately (as compared with older algorithms)
    • SM16 Metric - performance metric for Algorithm SM-15. It is a square root of the average of prediction differences squared expressed in percentage points (lower number means better recall prediction accuracy)
    • SM18 Metric - performance metric for Algorithm SM-18. It is computed in the same way as the analogous parameter on the SM16 Metric tab. Ideally, numbers on this tab should be lower though, indicating superiority of Algorithm SM-18
    • R-Metric - the ultimate performance metric for Algorithm SM-18. It compares it with the old Algorithm SM-15. It is the performance metric difference between the two metrics from previous tabs: SM16 Metric minus SM18 Metric. The higher the number, the greater the superiority of Algorithm SM-18 for a given collection. On days with a negative metric, the old algorithm would perform better than Algorithm SM-18. As forgetting is random, the actual metric outcome for a day is a lottery. Only a larger number of repetitions per day provide for consistent outcomes. For an exemplary graph based on 8800 data points, see: example below.
    • Topic load - proportion of topics scheduled for review on a given day among all elements scheduled. Topic load = Outstanding topics / outstanding elements. Topic load is expressed in percent for any given day. Topic load should not be confused with proportion of topics in sorting criteria. Proportion of topics is determined independent of the topic load if possible (e.g. for zero topic load, you cannot achieve proportion greater than zero, etc.). If you do not use auto-sorting, topic load tells you what proportion of your repetitions is made of topics.
  8. Overload - the number of executed repetitions divided by the number of outstanding repetitions (i.e. the percent of the actually executed repetitions)
    • Elements - the percent of all outstanding repetitions executed
    • Items - the percent of outstanding item repetitions executed
    • Topics - the percent of outstanding topic reviews done
  9. Priority protection - your processing capacity for high priority material on a given day. If your graph oscillates around priority of 3%, you will know that only top 3% of your learning material is guaranteed a timely repetition. You can increase that number by doing more work, reducing inflow of new material, deprioritizing less important elements, or reducing the randomization degree for the outstanding queue sorting criteria. Read about this important parameter in Prioritization Rulebook
    • Items - the highest priority item (with the lowest %) that was missed in repetitions. (The actual % in the outstanding queue can be much higher (because you mostly review high priorities))
    • Topics - the highest priority topic (with the lowest %) that was missed in repetitions. (see: exemplary graph below)

Use context menu

  • Next (Ctrl+Tab) - switch to the next major Use statistic tab
  • Previous (Shift+Ctrl+Tab) - switch back to the previous major Use statistic tab
  • Next secondary (Tab) - switch to the next tab in the secondary row
  • Previous secondary (Shift+Tab) - switch to the previous tab in the secondary row
  • Cut off left (Ctrl+Right arrow) - cut off the left part of the graph to zoom in on the rest of it
  • Cut off right (Ctrl+Left arrow) - cut off the right part of the graph to zoom in on the rest of it
  • Cut off top (Ctrl+Down arrow) - cut off the top part of the graph to zoom in on the rest of it
  • Export - export the currently displayed statistic to a CSV file (the file can be found in the collection's [STATS] subfolder
  • Import - import data for the currently displayed statistic
  • Close (Esc) - close the Analysis window

Examples

Use : Efficiency : R-Metric

SuperMemo: Algorithm SM-18 performance metric

Figure: R-Metric graph demonstrates superiority of Algorithm SM-18 over the old Algorithm SM-15 for the presented collection used in the testing period of full 4 years dating back to Apr 2, 2015. It was plotted using 24,104 data points (i.e. repetition cases with data from both algorithms), and smoothed up to show the trends. Multiple spots beneath the line of 0 at the vertical axis (R-metric<0) have been smoothed out (they correspond with days when the previous version of the algorithm appeared superior in a smaller sample of repetitions). Some positive and negative trends correspond with changes in the algorithm as data were collected in the new algorithm's testing period. A gradual increase in the metric in the months Feb-May 2016, might be a statistical aberration, or it might be the result of new interval values and a bigger R-metric for intervals departing from the optimum used in earlier SuperMemos. The latter interpretation might suggest that the benefits of Algorithm SM-18 can gradually increase over time

Use : Element count : Topics and tasks

SuperMemo: Toolkit : Statistics : Analysis : Use : Element count : Topics and tasks shows you how many topics and tasks there were in your collection on individual days

Figure: Exemplary graph showing the total number of topics stored in the collection. The data displayed at the bottom depends on the position of the mouse. In this example, it indicates that on Mar 6, 2009, there were 251,228 topics in the incremental reading process.

Use : Newly memorized

SuperMemo: Toolkit : Statistics : Analysis : Use : Newly memorized shows you how many elements you memorized on individual days

Figure: Exemplary graph showing the number of elements memorized on individual days. By sweeping away the left and right part of the graph, it is possible to zoom in on a small section of the learning process. The period shown on the graph spans from Jul 1, 2008 to Mar 19, 2009.

Use : Efficiency : Forgetting index

SuperMemo: Toolkit : Statistics : Analysis : Use : Efficiency : Forgetting index shows you the changes to the measured forgetting index on individual days

Figure: Exemplary graph enabling a more meaningful analysis of the forgetting index. Changes to forgetting index in Analysis use the daily measured forgetting index (previously: less informative cumulative measured forgetting index value was taken for the entire period since the last use of Toolkit : Statistics : Reset parameters : Forgetting index record). Note that the priority queue may distort the actual retention in your collection as measured values are primarily taken from top-priority material. Thus measured forgetting index should be understood as "forgetting index measured at repetitions", not as "overall measured forgetting index".

Use : Priority protection : Items

SuperMemo: Toolkit : Statistics : Analysis : Use : Priority protection : Items shows your actual processing capacity for high-priority items on individual days

Figure: Exemplary graph showing the highest priority item (with the lowest priority %) that was missed in repetitions on a given day. In the picture, you can see the increase in the item priority protection from 0.5% to 11% within a month. This is possible by focusing on honest prioritization, doing lots of repetitions, avoiding mass up-prioritizations, etc.

Use : Priority protection : Topics

SuperMemo: Toolkit : Statistics : Analysis : Use : Priority protection : Topics shows your actual processing capacity for high-priority topics on individual days

Figure: Exemplary graph showing the highest priority topic (with the lowest priority %) that was missed in reviews. The data displayed at the bottom depends on the position of the mouse. In this example, it indicates that on Jun 26, 2013, the worst miss amongst topics was a topic with priority equal to 7.613% of the queue of the most important topics to review.

Forgetting

400 forgetting curves are independently plotted in order to compute the SInc matrix in Algorithm SM-18. These correspond to 20 stability categories and 20 difficulty categories. By choosing a proper combination of tabs at the bottom of the graph, you can select a forgetting curve of interest. Horizontal axis represents time expressed as: (1) retrievability, or (2) days (only for the first repetition). Vertical axis represents knowledge retention in percent.

SuperMemo: The first forgetting curve (first repetition for items with no lapses)

Figure: The first forgetting curve for newly learned knowledge collected with SuperMemo. Power approximation is used in this case due to the heterogeneity of the learning material freshly introduced in the learning process. Lack of separation by memory complexity results in superposition of exponential forgetting with different decay constants. On a semi-log graph, the power regression curve is logarithmic (in yellow), and appearing almost straight. The curve shows that in the presented case recall drops merely to 58% in four years, which can be explained by a high reuse of memorized knowledge in real life. The first optimum interval for review at retrievability of 90% is 3.96 days. The forgetting curve can be described with the formula R=0.9906*power(interval,-0.07), where 0.9906 is the recall after one day, while -0.07 is the decay constant. In this is case, the formula yields 90% recall after 4 days. 80,399 repetition cases were used to plot the presented graph. Steeper drop in recall will occur if the material contains a higher proportion of difficult knowledge (esp. poorly formulated knowledge), or in new students with lesser mnemonic skills. Curve irregularity at intervals 15-20 comes from a smaller sample of repetitions (later interval categories on a log scale encompass a wider range of intervals).

Forgetting (UF)

This tab displays forgetting data as used in older SuperMemos. It is only used to record and display comparisons between the new algorithm and the older algorithms.

400 forgetting curves are independently plotted in order to compute the RF matrix in Algorithm SM-15. These correspond to 20 repetition number categories and 20 A-Factor categories. For data representation convenience, the columns of the RF matrix for the first repetition are indexed by the number of memory lapses rather than by the A-Factor. By choosing a proper combination of tabs at the bottom of the graph, you can select a forgetting curve of interest. Horizontal axis represents time expressed as: (1) U-Factor, i.e. the ratio of subsequent inter-repetition intervals, or (2) days (only for the first repetition). Vertical axis represents knowledge retention in percent.

SuperMemo: Toolkit : Statistics : Analysis : Forgetting (UF) graphs for 20 repetition number categories multiplied by 20 A-Factor categories

Figure: Toolkit : Statistics : Analysis : Forgetting (UF) for 20 repetition number categories multiplied by 20 A-Factor categories. In the picture, blue circles represent data collected during repetitions. The larger the circle, the greater the number of repetitions recorded. The red curve corresponds with the best-fit forgetting curve obtained by exponential regression. For ill-structured material the forgetting curve is crooked, i.e. not exactly exponential. The horizontal aqua line corresponds with the requested forgetting index, while the vertical green line shows the moment in time in which the approximated forgetting curve intersects with the requested forgetting index line. This moment in time determines the value of the relevant R-Factor, and indirectly, the value of the optimum interval. For the first repetition, R-Factor corresponds with the first optimum interval. The values of O-Factor and R-Factor are displayed at the top of the graph. They are followed by the number of repetition cases used to plot the graph (i.e. 21,303). At the beginning of the learning process, there is no repetition history and no repetition data to compute R-Factors. It will take some time before your first forgetting curves are plotted. For that reason, the initial value of the RF matrix is taken from the model of a less-than-average student. The model of average student is not used because the convergence from poorer student parameters upwards is faster than the convergence in the opposite direction. The Deviation parameter displayed at the top tells you how well the negatively exponential curve fits the data. The lesser the deviation, the better the fit. The deviation is computed as a square root of the average of squared differences (as used in the method of least squares).

SuperMemo: In SuperMemo 18, forgetting curves can be normalized over A-Factors at different repetition categories. As a result, you can display (1) your cumulative forgetting curve (blue dots), (2) its negative exponential approximation used by SuperMemo (red line) and (3) individual data points from all forgetting curves collected by SuperMemo (yellow circles).

Figure: In SuperMemo 18, forgetting curves can be normalized over A-Factors (All AFs) at different repetition categories (All Reps)(as in the picture above). As a result, you can display:

  1. your cumulative forgetting curve (blue dots)
  2. its negative exponential approximation used by SuperMemo (yellow line)
  3. individual data points from all forgetting curves collected by SuperMemo (yellow circles with red border).

Time (normalized) stands at 100% for maximum U-Factor for each entry of the RF matrix. Decay stands for the decay constant of the forgetting curve. Cases shows the number of repetition cases used in plotting the graph. Repetitions stands for the number of repetitions executed when determining R-Factors. Deviation is a square root of the average of squared differences between the data and the approximated forgetting curve.

Stabilization

The tab marked Stabiliz. in Analysis can be used to inspect stabilization curves in SuperMemo.

Inspecting stabilization curves for various levels of stability and difficulty

Figure: SuperMemo 18 makes it possible to inspect 400 independent stabilization curves. These curves differ by levels of stability and difficulty. The picture shows the curve for difficulty=0.58 and stability=59 days. The horizontal axis shows retrievability (as percent). The vertical axis shows stabilization (as a ratio of stability). Early repetitions bring little stabilization, and the effect of review increases rapidly with a drop to retrievability

Graphs

Forgetting Index vs. Priority

Forgetting Index vs. Priority - this graph shows you how forgetting depends on the priority of the learning material. If you always go through your outstanding material in time and if you stick to the 20 rules of knowledge formulation, your measured forgetting index should be the same as the requested forgetting index. This graph will then show a flat straight line passing horizontally at the level of the forgetting index. However, once you start overloading the learning process and using Postpone or Mercy, your measured forgetting index will increase. If you always sort your repetitions by priority (e.g. by leaving Learn : Sorting : Auto-sort repetitions checked), you should still ensure that your measured forgetting index equals the requested forgetting index for high priority elements (i.e. on the left side of the graph). Your measured forgetting index for low priority material will be higher, and the greater the overload the greater the difference. If you change your repetition sorting criteria by increasing randomization of repetitions, this graph will be flatter and the measured forgetting index for high priority material will be higher.

First interval

First interval - the length of the first interval after the first repetition depends on the number of times a given item has been forgotten. Note that the first repetition here means the first repetition after forgetting, not the first repetition ever. In other words, a twice repeated item will have the repetition number equal to one after it has been forgotten; the repetition number will not equal three. The first interval graph shows exponential regression curve that approximates the length of the first interval for different numbers of memory lapses (including the zero-lapses category that corresponds with newly memorized items). In the graph below, blue circles correspond to data collected in the learning process (the greater the circle, the more repetitions have been recorded).

SuperMemo: Toolkit : Statistics : Analysis : Graphs : First Interval shows you exponential regression curve that approximates the length of the first interval for different numbers of memory lapses

Figure: In the graph above, which includes data from over 130,000 repetitions, newly memorized items are optimally repeated after seven days. However, items that have been forgotten 10 times (which is rare in SuperMemo) will require an interval of two days. (Due to logarithmic scaling, the size of the circle is not linearly proportional to the data sample; the number of repetition cases for lapses=0 is by far larger than for lapses=10, as can be seen in Toolkit : Statistics : Analysis : Distributions : Lapses)

D-Factor vs. A-Factor

D-Factor vs. A-Factor - DF-AF graph shows decay constants of the power approximation of R-Factors along columns of the RF matrix. You need to understand SuperMemo Algorithm SM-15 to understand this graph. The horizontal axis represents A-Factor, while the vertical axis represents D-Factor (i.e. Decay Factor). D-Factor is a decay constant of the power approximation of curves that can be inspected in the Approximations tab of the Analysis dialog box.

First Grade vs. A-Factor

First Grade vs. A-Factor - G-AF graph correlates the first grade obtained by an item with the ultimate estimation of its A-Factor value. At each repetition, the current element's old A-Factor estimation is removed from the graph and the new estimation is added. This graph is used by SuperMemo Algorithm SM-15 to quickly estimate the first value of A-Factor at the moment when all we know about an element is the first grade it has scored in its first repetition.

Grade vs. Forgetting Index

Grade vs. Forgetting Index - FI-G graph correlates the expected forgetting index with the grade scored at repetitions. You need to understand SuperMemo Algorithm SM-15 to understand this graph. You can imagine that the forgetting curve graph might use the average grade instead of the retention on its vertical axis. If you correlated this grade with the forgetting index, you would arrive at the FI-G graph. This graph is used to compute an estimated forgetting index that is in turn used to normalize grades (for delayed or advanced repetitions) and estimate the new value of item's A-Factor. The grade is computed using the formula: Grade=Exp(A*FI+B), where A and B are parameters of an exponential regression run over raw data collected during repetitions.

SuperMemo: Toolkit : Statistics : Analysis : Graphs : Grade vs. Forgetting index graph

Distributions

Intervals

Intervals - distribution of inter-repetition intervals in a given collection. By changing the setting of the combo box at the bottom of the dialog, you can see interval distributions for all elements in the collection, for only items, and for only topics.

A-Factors

A-Factors - distribution of A-Factors in a given collection. The distribution itself is not used in Algorithm SM-15, and merely results from it. Note that topics and items use different A-Factor categories. Choose Items only or Topics only at the bottom to see distributions for items or topics respectively. If you choose All elements you will see a less meaningful conglomerate distribution indexed by A-Factor category number.

SuperMemo: Toolkit : Statistics : Analysis : Distributions : A-Factors shows you how A-Factors are distributed in your collection

Difficulties

Difficulties - distribution of item difficulties as determined by Algorithm SM-18. See also: difficulties

Repetitions

Repetitions - distribution of the number of repetitions of items or topics in a given collection; only memorized elements are considered in the distribution, i.e. there is no zero-repetitions category.

Lapses

Lapses - distribution of the number of times particular items in the collection have been forgotten; only memorized elements are considered. You should never have more than 10% of items with more than 3-4 lapses. If this is not the case, you should re-examine the way in which you formulate items (see 20 Rules of Formulating Knowledge). As topics are never "forgotten" at repetition time, there are no lapses listed in Lapses if you select Topics only in the element type combo box.

SuperMemo: Toolkit : Statistics : Analysis : Distributions : Lapses shows you how memory lapses are distributed in your collection

3-D Curves

These graphs show the forgetting curves in 3 dimensions. There are 20 graphs available with A-Factor tabs at the bottom. Each graph corresponds to a different A-Factor (from 1.2 to 6.9). The X axis represents time (as in Forgetting curves). The Y axis represents repetition category (roughly corresponding to the repetition number). The vertical Z axis represents retention in percent (i.e. how much is remembered at a given moment in time). For better visibility, you can rotate 3-D forgetting curves with Rotate, or smooth them with Smooth. Rotate swings the graph on its Z axis by 90 degrees clockwise (as viewed from above). This means that X and Y axes get swapped. Note that smoothing proceeds along the Y axis. This makes it possible to align incompletely plotted forgetting curves from different repetition categories. However, once you click Rotate, Y axis smoothing will produce a different result (e.g. after the first Rotate, smoothing will reduce kinks along the repetition categories). For the best illustration of the forgetting process, smooth the graph before you rotate it.

SuperMemo: Exemplary 3-D graph of forgetting curves for A-Factor=3.6

Figure: Exemplary 3-D graph of forgetting curves for A-Factor=3.6. Axes: X - repetition category, Y - recall in percent (knowledge retention), Z (the incoming axis) - time (expressed by U-Factor). The graph has been smoothed along the Y-axis (by averaging results corresponding to different repetition categories), and rotated once for better visibility (the slope illustrates forgetting in time that is roughly similar for various repetition categories except those with high repetition number that may reflect scarcity of data for long intervals). For a better view of the same data in 3D see: Toolkit : Memory : 4D graphs : Forgetting.

3-D Graphs

3-D graphs that visually represent RF, OF, and Cases matrices (see: Matrices below)

S. decay

Graphs that show how SInc slope is computed.

OF decay

20 power approximation curves that show the decline of R-Factors along columns of the RF matrix. You need to understand SuperMemo Algorithm SM-15 to understand these graphs. For each A-Factor, with increasing values of the repetition number, the value of R-Factor decreases (at least theoretically it should decrease). Power regression is used to illustrate the degree of this decline that is best reflected by the decay constant called here D-Factor. By choosing the A-Factor tab at the bottom of the graph, you can view a corresponding R-Factor approximation curve. The horizontal axis represents the repetition number category, while the vertical axis represents R-Factor. The value of D-Factor is shown at the top of the graph. The blue polyline shows R-Factors as derived from repetition data. The red curve shows the fixed-point power approximation of R-Factor. The green curve shows the fixed-point power approximation of R-Factor derived from the OF matrix; this is equivalent to substituting the D-Factor obtained by fixed-point power approximation of R-Factors for the D-Factor obtained from DF-AF linear regression. A fixed-point approach is used for both approximations since for the repetition number equal to two, R-Factor equals A-Factor.

Matrices

You need to understand SuperMemo Algorithm to understand these tables.

SuperMemo: Toolkit : Statistics : Matrices : Intervals displays the matrix of optimum intervals derived from the OF matrix

FAQ

Further reading