I created Reasoner-thinking with Rule For AI

I created Reasoner-thinking with Rule For AI. The only limitation is the context memory of the models and the limited length of the Cursor for the Rule For AI file.

IX. DETAILED RULES FOR REASONING METHODS:

  1. GENERAL INTRODUCTION

This document describes in detail the reasoning methods, their uses, and response structures. The four main methods presented are: MCTS Reasoning, Beam Search Reasoning, R1 Reasoning, and Hybrid Reasoning. Each method has its own procedure, response structure, and is suitable for specific types of problems.

  1. MCTS REASONING [MCTS]

2.1. Description

Monte Carlo Tree Search is an inference method based on building decision trees and random simulation to find the optimal branch. Suitable for problems with large state spaces, requiring a balance between exploration and exploitation.

2.2. Phased process

Phase 1: Initiation - Identify the problem, state, actions and goals

Phase 2: Selection - Select the prospect button using UCB

Phase 3: Expand - Expand the selected node with possible actions

Phase 4: Simulation - Run the simulation from the new state to the final state

Phase 5: Update - Update statistical information for nodes in the path

Phase 6: Decision - Choose the best action based on tree statistics

2.3. Feedback structure

[MCTS Reasoning]

  • reasoningStage: <stage/6>

  • nodeID: “”

  • nodeDepth:

  • thought: “”

  • simulationsCompleted: <number/total>

  • explorationConstant:

  • topPaths: [

{path: “”, visits: , winRate: <x.■■>, ucbValue: <x.■■>},

{path: “”, visits: , winRate: <x.■■>, ucbValue: <x.■■>}

]

  • currentStrategy: “<exploration/exploitation>”

  • unexploredActions:

  • confidence: <0-100%>

  • nextStepNeeded: <true/false>

  • completedChecklist: [“<item 1>”, “<item 2>”, …]

  • pendingChecklist: [“<item 1>”, “<item 2>”, …]

2.4. Usage examples

When required: [Deduction-MCTS] Analyze optimal strategies for the traveling agent problem

[MCTS Reasoning]

  • reasoningStage: 2/6

  • nodeID: “A→B→D”

  • nodeDepth: 3

  • thought: “Applying UCB to select the next node from current state A→B→D. The child node with the highest UCB value is D→C with UCB=1.42 based on high success rate 0.65 and low exploration level (15 visits).”

  • simulationsCompleted: 48/100

  • explorationConstant: 1.41

  • topPaths: [

{path: “A→B→D→C→E→A”, visits: 15, winRate: 0.65, ucbValue: 1.42},

{path: “A→B→D→E→C→A”, visits: 12, winRate: 0.58, ucbValue: 1.38}

]

  • currentStrategy: “Explore”

  • unexploredActions: 2

  • confidence: 52%

  • nextStepNeeded: true

  • completedChecklist: [“Determine current state”, “Calculate UCB for child nodes”, “Select most promising node”]

  • pendingChecklist: [“Expand D→C node”, “Run simulation from new state”]

2.5. Optimization for MCTS Reasoning

  • Adaptive exploration constant: Adjust the UCB exploration constant according to the process
  • Formula: C = Base + (Max-Base) * (1 - CurrentDepth/MaxDepth)^2

  • Start high (exploration) and decrease (exploitation) with depth

  • Progressive widening: Limit the number of expanding branches according to the formula
  • Number of branches = ⌊ kN^ α ⌋ , with k=1.5, α=0.4, N is the number of node visits
  • Rapid Action Value Estimation (RAVE): Uses information from similar nodes
  • Combine UCB value with RAVE statistic in decreasing weight
  • Backpropagation with decay: Backpropagate values with decay coefficient according to depth
  • Value = Current Value * (1-α) + New Value * α * Îł^d

2.6. Mandatory Checklist for Each MCTS Stage

Phase 1 (Initialization):

  • Clearly define the initial state

  • Define all possible actions

  • Define the state evaluation function

  • Set search depth limit

  • Determine the termination condition

Phase 2 (Optional):

  • Calculate UCB value for every child node

  • Determine current exploration/exploitation strategy

  • Select the button with the highest UCB

  • Record the path to the selected node

Phase 3 (Expansion):

  • Determine all possible actions from the selected button

  • Apply Progressive widening technique if necessary

  • Create new child button for each action

  • Initialize statistics for new node

Phase 4 (Simulation):

  • Choose the right simulation strategy

  • Perform simulation from new state to final state

  • Evaluation of simulation results

  • Record the simulation path

Phase 5 (Update):

  • Update the number of visits for each node in the path

  • Update win rate for each node in the path

  • Apply Backpropagation with decay technique if necessary

  • Update RAVE information if used

Phase 6 (Decision):

  • Evaluate all child nodes of the root node

  • Compare based on number of visits or win rate

  • Choose the best action

  • Evaluate the reliability of the decision

2.7. MCTS Phase Transition Prerequisites

Conditions for transitioning from Phase 1 to Phase 2:

  • Fully defined initial state

  • Fully identified all possible actions

  • Status evaluation function set up

Conditions for transitioning to Phase 2→3:

  • Selected the most promising node with the highest UCB

  • Fully recorded path to selected node

Conditions for transitioning to Phase 3→4:

  • Created at least one new child node

  • Progressive widening technique applied if necessary

Conditions for transitioning to Stage 4→5:

  • Completed at least one simulation

  • Simulation results evaluated

Conditions for transitioning to Stage 5→6:

  • Updated statistics for all nodes in the path

  • The required number of simulations has been achieved OR

  • Minimum confidence threshold has been reached (usually 75%)

Stage 6 End Conditions:

  • The optimal action has been selected.

  • Reliability is at least 80%

  • Clearly explained the reason for choosing this action

  1. BEAM SEARCH REASONING [BEAM]

3.1. Description

Beam Search maintains a limited set of best solutions (beams) at each step and expands them, eliminating poorer solutions. Suitable for ordered search problems, optimization problems with multiple choices.

3.2. Phased process

Phase 1: Initialization - Define the problem, evaluation function and beam width

Phase 2: Expand - Expand each current ray with all possible actions

Phase 3: Evaluation - Evaluate all new states according to the evaluation function

Phase 4: Pruning - Keep only the best condition to continue

Phase 5: Check - Check for termination or convergence conditions

3.3. Feedback structure

[BEAM Reasoning]

  • reasoningStage: <stage/5>

  • thought: “”

  • beamWidth:

  • activeBeams:

  • expansionStep:

  • currentBeams: [

{path: “”, score: <x.■■>, features: “”},

{path: “”, score: <x.■■>, features: “”}

]

  • recentlyPruned: {

count: ,

topPruned: “”,

reason: “”

}

  • diversityMetric: <0-1>

  • convergenceStatus: “”

  • confidence: <0-100%>

  • nextStepNeeded: <true/false>

  • completedChecklist: [“<item 1>”, “<item 2>”, …]

  • pendingChecklist: [“<item 1>”, “<item 2>”, …]

3.4. Usage examples

When asked: [Deduction-BEAM] Analyze how to rank the top 5 candidates for a management position

[BEAM Reasoning]

  • reasoningStage: 3/5

  • thought: “Currently evaluating all 15 candidate profiles based on 5 criteria: experience (0.3), leadership skills (0.25), qualifications (0.15), communication skills (0.2), and innovation (0.1). The evaluation score is calculated by the sum of the criteria with weights.”

  • beamWidth: 5

  • activeBeams: 5

  • expansionStep: 2

  • currentBeams: [

{path: “Candidate A”, score: 8.7, features: “10+ years of experience, MBA, good communication”},

{path: “Candidate C”, score: 8.5, features: “8 years experience, excellent leadership”}

]

  • recentlyPruned: {

count: 3,

topPruned: “Candidate F”,

reason: “Lack of leadership experience (2.5/10)”

}

  • diversityMetric: 0.72

  • convergenceStatus: “70% - Top 3/5 positions stable after 2 steps”

  • confidence: 65%

  • nextStepNeeded: true

  • completedChecklist: [“Define evaluation criteria”, “Set weights”, “Evaluate all candidates”]

  • pendingChecklist: [“Pruning candidate list”, “Checking convergence”]

3.5. Optimizing for Beam Search

  • Adaptive beam width: Adjust beam width according to processing stage
  • Initial stage: Broad beam (50-100% initialization)

  • Middle stage: Medium beam (30-70% initialization)

  • Final stage: Narrow beam (10-50% initialization)

  • Diversity enforcement: Ensures diversity in rays is maintained
  • Penalize too similar solutions using a distance function

  • Keep only solutions with distance > threshold

  • Priority scheduling: Schedule expansion by priority
  • High priority: Promising (high score) and different rays

  • Medium Priority: Promising but similar beam

  • Low priority: Less promising beam

  • Smart pruning: Smart pruning based on multiple criteria
  • Not only based on scores but also on development potential

  • Use Bloom filters to avoid duplicate states

3.6. Mandatory Checklist for Each Beam Phase

Phase 1 (Initialization):

  • Identify the problem to be solved

  • Definition of evaluation function and criteria

  • Determine the weight for each criterion

  • Set beam width

  • Initialize the initial ray

Phase 2 (Expansion):

  • Identify all possible actions for each ray

  • Create new state list from each ray

  • Apply Adaptive beam width if needed

  • Ensure sufficient number of states are expanded

Phase 3 (Assessment):

  • Apply the evaluation function to every new state

  • Calculate score for each state

  • Apply Diversity enforcement if necessary

  • Rate all new statuses

Phase 4 (Pruning):

  • Apply Smart pruning

  • Remove below threshold state

  • Save the best status

  • Calculate the diversity of retained rays

Phase 5 (Testing):

  • Check the end condition

  • Evaluate the convergence of the solution

  • Determine the best solution

  • Evaluate the reliability of the results

3.7. Beam Phase Transition Prerequisites

Conditions for transitioning from Phase 1 to 2:

  • Fully defined evaluation function with weights

  • Set the appropriate beam width

  • Initial ray initialized validly

Conditions for transitioning to Phase 2→3:

  • Expanded each ray with at least one possible action

  • Created at least beamWidth x 2 new states to evaluate

Conditions for transitioning to Phase 3→4:

  • Rated 100% new condition

  • Scored for all statuses

  • Ranked states by score

Conditions for transitioning to Stage 4→5:

  • Saved correctly k best state

  • Removed below threshold states

  • Calculated new ray diversity

Phase 5 End Conditions:

  • Has met one of the following conditions:
  • Convergence to solution (top rays are stable)

  • Achieve maximum number of expansion steps

  • Reliability threshold ≥ 85%

  • The best solution has been determined
  1. R1 REASONING [R1]

4.1. Description

R1 Reasoning is a comprehensive transformer-based analysis method that focuses on analyzing each component of a problem and synthesizing them into a consistent assessment. Suitable for complex analysis, evaluation, and synthesis.

4.2. Phased process

Phase 1: Problem Analysis - Analyze and decompose the problem into its components

Phase 2: Component Analysis - In-depth analysis of each component

Phase 3: Synthesis - Synthesize analysis from all components

Phase 4: Evaluation - Evaluate for consistency and comprehensiveness

Phase 5: Conclusion - Draw conclusions and recommendations

4.3. Feedback structure

[R1 Reasoning]

  • reasoningStage: <stage/5>

  • thought: “”

  • components: [

{name: “”, status: “<analyzed/not>”, depth: <1-5>},

{name: “”, status: “<analyzed/not>”, depth: <1-5>}

]

  • currentComponent: “”

  • keyInsights: [“<insight 1>”, “<insight 2>”]

  • contradictions: [

{elements: [“”, “”], resolution: “<resolved/unresolved>”}

]

  • hypotheses: {

primary: “”,

alternatives: [“<alternative 1>”, “<alternative 2>”]

}

  • integrationScore: <0-100%>

  • consistency: <0-100%>

  • confidence: <0-100%>

  • nextStepNeeded: <true/false>

  • completedChecklist: [“<item 1>”, “<item 2>”, …]

  • pendingChecklist: [“<item 1>”, “<item 2>”, …]

4.4. Usage examples

When asked: [Reasoning-R1] Analyze the feasibility of software development project X

[R1 Reasoning]

  • reasoningStage: 2/5

  • thought: “Analyzing the technical aspects of project X in detail. In terms of backend technology, the proposal to use Node.js has advantages in performance and scalability, but has security problems when handling sensitive data. Additional security measures should be considered.”

  • components: [

{name: “Technical aspects”, status: “under analysis”, depth: 4},

{name: “Financial aspects”, status: “analyzed”, depth: 3},

{name: “Human resources aspect”, status: “not analyzed”, depth: 0}

]

  • currentComponent: “Technical aspect - Technology stack”

  • keyInsights: [“Estimated implementation time 8 months”, “Requires 4 fullstack developers”]

  • contradictions: [

{elements: [“High security requirements”, “Time pressure”], resolution: “Security priority recommended”}

]

  • hypotheses: {

primary: “Project feasible with time adjustment”,

alternatives: [“Need more manpower”, “Adjust project scope”]

}

  • integrationScore: 65%

  • consistency: 78%

  • confidence: 72%

  • nextStepNeeded: true

  • completedChecklist: [“Financial aspects analysis”, “Preliminary technical aspects analysis”]

  • pendingChecklist: [“Detailed technology stack analysis”, “Human resource aspect analysis”]

4.5. Optimization for R1 Reasoning

  • Component prioritization: Prioritize analysis of important components first
  • Determine importance by contextual analysis

  • Allocate time in proportion to importance

  • Dynamic depth adjustment: Adjust the analysis depth according to needs
  • Key Ingredients: Depth 4-5

  • Sub-component: Depth 2-3

  • Secondary component: Depth 1-2

  • Contradiction-focused analysis: Focuses on potential points of conflict
  • Identify potentially conflicting components prior to analysis

  • Apply further analysis to these points

  • Insight accumulation: Accumulate insights over time
  • Store and evaluate the importance of each insight

  • Prioritize development based on key insights

4.6. Mandatory Checklist for Each R1 Phase

Phase 1 (Problem Analysis):

  • Identify the overall problem to be analyzed

  • Decompose the problem into at least 3 main components

  • Determine the relationship between components

  • Evaluate the importance of each component

  • Set the required depth of analysis for each component

Phase 2 (Component Analysis):

  • Analyze each component to the specified depth

  • Identify at least 2 key insights for each component

  • Detect potential conflicts within each component

  • Propose hypotheses for each component

  • Evaluate the reliability of the analysis for each component

Phase 3 (Summary):

  • Combine analysis from all components

  • Identify at least 3 relationships between components

  • Detect and resolve conflicts between components

  • Building a logical synthesis framework

  • Evaluate the comprehensiveness of the synthesis

Phase 4 (Assessment):

  • Evaluate the consistency of the overall analysis

  • Identify at least 2 strengths and 2 weaknesses

  • Assess the coverage of the analysis

  • Test the validity of hypotheses

  • Overall reliability rating

Phase 5 (Conclusion):

  • Draw clear conclusions based on analysis

  • Propose at least 3 specific recommendations

  • Identify the limitations of the analysis

  • Propose further research/analysis directions

  • Summarize the most important insights

4.7. Prerequisites for phase transition R1

Conditions for transitioning from Phase 1 to 2:

  • Identified at least 3 problem components

  • Established relationships between components

  • Determined the level of importance and depth required

Conditions for transitioning to Phase 2→3:

  • At least 80% of the ingredients have been analyzed

  • Each component has at least 1 important understanding

  • Potential conflicts have been identified and recorded.

Conditions for transitioning to Phase 3→4:

  • Synthesized analysis from all components

  • Resolved all detected conflicts

  • Achieved minimum consistency of 70%

Conditions for transitioning to Stage 4→5:

  • Completed overall assessment

  • Clearly identified strengths and weaknesses

  • Achieved overall reliability of at least 75%

Phase 5 End Conditions:

  • Has drawn a clear conclusion

  • Proposed at least 3 specific recommendations

  • Summarized the most important insights

  • Reliability is at least 80%

  1. HYBRID REASONING [HYBRID]

5.1. Description

Hybrid Reasoning combines different methods to take advantage of the strengths of each. Suitable for complex, multidimensional problems that have both random and structured elements.

5.2. Phased process

Phase 1: General Analysis - Use R1 to analyze the overall problem

Phase 2: Stochastic Analysis - Using MCTS for Uncertainty Components

Phase 3: Parallel Analysis - Using Beam Search to Consider Multiple Solutions

Phase 4: Integration - Combining results from different methods

Phase 5: Conclusion - General summary and evaluation

5.3. Feedback structure

[HYBRID Reasoning]

  • reasoningStage: <stage/5>

  • thought: “”

  • activeMethod: “<R1/MCTS/BEAM>”

  • methodDistribution: {

r1: ,

mcts: ,

beam:

}

  • componentMapping: [

{component: “”, method: “”, reason: “”},

{component: “”, method: “”, reason: “”}

]

  • currentFocus: “<current component/aspect>”

  • methodMetrics: {

r1: {completeness: <0-100%>, insights: },

mcts: {simulations: , confidence: <0-100%>},

beam: {beams: , bestScore: <x.â– â– >}

}

  • integrationChallenges: [“<challenge 1>”, “<challenge 2>”]

  • resolutionApproach: “”

  • overallConfidence: <0-100%>

  • nextStepNeeded: <true/false>

  • completedChecklist: [“<item 1>”, “<item 2>”, …]

  • pendingChecklist: [“<item 1>”, “<item 2>”, …]

5.4. Usage examples

When required: [Inference-HYBRID] Investment strategy analysis for diversified portfolio

[HYBRID Reasoning]

  • reasoningStage: 3/5

  • thought: “Currently using Beam Search to evaluate 5 different asset allocations in parallel. Each is evaluated based on expected return, risk (standard deviation), correlation, and liquidity. The 40% stocks, 30% bonds, 20% real estate, 10% cash strategy currently has the highest score.”

  • activeMethod: “BEAM”

  • methodDistribution: {

r1: 35%,

mcts: 40%,

beam: 25%

}

  • componentMapping: [

{component: “Market risk analysis”, method: “R1”, reason: “Structural factors, need in-depth analysis”},

{component: “Profit Simulation”, method: “MCTS”, reason: “Random factor, multiple scenarios”}

]

  • currentFocus: “Compare asset allocation options”

  • methodMetrics: {

r1: {completeness: 85%, insights: 7},

mcts: {simulations: 500, confidence: 72%},

beam: {beams: 5, bestScore: 8.6}

}

  • integrationChallenges: [“Contradiction between MCTS and R1 results on market risk”]

  • resolutionApproach: “Sensitivity analysis for different market scenarios”

  • overallConfidence: 76%

  • nextStepNeeded: true

  • completedChecklist: [“Market Risk Analysis (R1)”, “Profit Simulation (MCTS)”, “Create 5 Allocation Options”]

  • pendingChecklist: [“Fully evaluate options”, “Resolve risk conflicts”]

5.5. Optimization for Hybrid Reasoning

  • Method switching criteria: Method switching criteria according to the nature of the problem
  • Increased complexity → Moving from single method to hybrid

  • Detect clear logical structure → Increase weight R1

  • Detect large state space → Increase MCTS weights

  • Need to compare multiple solutions → Increase Beam weight

  • Resource allocation: Allocate resources according to importance
  • 70% resources for the most complex component

  • 20% resources for the second most important component

  • 10% resources for remaining components

  • Synergy maximization: Maximize synergy between methods
  • Share intermediate results between methods

  • Combine strengths: MCTS for discovery, Beam for comparison, R1 for analysis

  • Adaptable confidence thresholds: Adaptable confidence thresholds
  • Important issue: High threshold (85-95%)

  • Medium problem: Medium threshold (75-85%)

  • Common problem: Low threshold (65-75%)

5.6. Mandatory Checklist for Each Hybrid Phase

Phase 1 (General Analysis):

  • Identify the problem to be solved

  • Decompose the problem into components

  • Determine the appropriate method for each component

  • Set up initial resource allocation

  • Determine the relationship between components

Phase 2 (Randomized Analysis):

  • Identify components that require MCTS

  • Apply MCTS to uncertain components

  • Run enough simulations

  • Identify promising solutions from MCTS

  • Evaluate the reliability of MCTS results

Phase 3 (Parallel Analysis):

  • Identify the components that need Beam Search

  • Create and evaluate parallel solutions

  • Apply Diversity enforcement

  • Prune ineffective solutions

  • Identify top solutions from Beam Search

Phase 4 (Integration):

  • Collect results from all methods

  • Detect and resolve conflicts

  • Create a consistent integration framework

  • Evaluate synergies between methods

  • Adjust method weights if needed

Phase 5 (Conclusion):

  • Evaluate integrated solutions

  • Determine overall reliability

  • Draw conclusions from all methods

  • Propose specific recommendations

  • Identify limitations and directions for improvement

5.7. Hybrid Phase Transition Prerequisites

Conditions for transitioning from Phase 1 to Phase 2:

  • Decomposed the problem into at least 3 components

  • Applied R1 analysis to the entire problem

  • Determined appropriate methods for each component

Conditions for transitioning to Phase 2→3:

  • MCTS has been applied to all uncertain components

  • Run at least 100 simulations for each MCTS component

  • Achieved minimum MCTS reliability of 70%

Conditions for transitioning to Phase 3→4:

  • Beam Search applied to all elements to be compared

  • Identified at least 3 promising solutions

  • All solutions were evaluated according to the same criteria

Conditions for transitioning to Stage 4→5:

  • Integrated results from all methods

  • Resolved all detected conflicts

  • Achieved minimum integration consistency of 75%

Phase 5 End Conditions:

  • A synthesis of conclusions from all methods has been drawn.

  • Clearly assessed the reliability of each component

  • Overall reliability is at least 80%

  • Provided at least 3 specific recommendations

  1. CONFLICT RESOLUTION MECHANISM

6.1. Conflict detection and resolution process

Phase 1: Detecting contradictions - Using the comparison and contrast method

Stage 2: Conflict Classification - Classification by Level, Type and Origin

Phase 3: Determine Weights - Assign weights to each data source and method

Stage 4: Conflict Resolution - Apply appropriate resolution strategies

Phase 5: Update Confidence - Adjust the confidence of the final result

6.2. Conflict resolution strategies

  • Majority â– â– â– â– â– â– â– â– â– â– : Choose the outcome that is supported by the majority of methods

  • Priority weight: Prioritize the method with higher reliability for the specific problem

  • Cause analysis: Identify the source of the conflict and resolve it.

  • Reanalyze with new parameters: Change the parameters and rerun the method

  • Conditional combination: Integrate results with conditions on the scope of application

6.3. Dynamic weights for methods

  • R1 Reasoning: High weighting for structural, logical and comprehensive analysis (0.6-0.9)

  • MCTS Reasoning: High weight for problems with random elements and many branches (0.7-0.9)

  • Beam Search: High weight for optimal search and solution comparison (0.6-0.8)

  • Hybrid: Weights adjusted according to each problem component (0.5-0.95)

6.4. Detailed structure of the conflict report

  • conflictDetected: <true/false>

  • conflictType: “”

  • conflictLocation: “<conflicting component/result>”

  • methodsInvolved: [“<method 1>”, “<method 2>”]

  • conflictSeverity: <0-10>

  • resolutionStrategy: “”

  • resolutionRationale: “”

  • confidenceAdjustment: <±%>

  • resolutionOutcome: “”

6.5. Mandatory checklist for conflict resolution

When a conflict is detected:

  • Clearly identify the nature of the conflict

  • Severity rating (1-10)

  • Classify the type of conflict (data, logic, approach)

  • Identify relevant methods/components

  • Record full information in the report

When resolving conflicts:

  • Determine appropriate solution strategy

  • Clearly state the reason for choosing this strategy

  • Apply strategy consistently

  • Evaluate the results after resolution

  • Update overall reliability

  1. METRICS FOR ASSESSING THE QUALITY OF REASONING

7.1. Overall evaluation system

  • Accuracy: The degree of agreement with the standard or actual result (0-100%)

  • Consistency: The degree of internal inconsistency (0-100%)

  • Coverage: The extent to which all aspects of the problem are covered (0-100%)

  • Explainability: Ability to explain reasoning steps (0-100%)

  • Efficiency: The ratio between the quality of the result and the resources used (0-100%)

  • Composite reliability: Composite from all metrics (0-100%)

7.2. Detailed index for each method

R1 Reasoning:

  • componentCoverage: <0-100%> - Coverage of the problem components

  • insightRelevance: <0-100%> - Relevance of insights

  • logicalCoherence: <0-100%> - Logical coherence in analysis

  • counterArgumentQuality: <0-100%> - Counterargument Quality

MCTS Reasoning:

  • explorationDepth: <0-100%> - State space exploration depth

  • stateSpaceCoverage: <0-100%> - State space coverage

  • simulationAccuracy: <0-100%> - Accuracy of simulations

  • pruningEfficiency: <0-100%> - Unpromising pruning efficiency

Beam Search:

  • diversityScore: <0-100%> - Diversity of solutions

  • optimalityGap: <0-100%> - Distance to optimal solution

  • pruningPrecision: <0-100%> - Pruning precision

  • convergenceRate: <0-100%> - Convergence rate to good solution

Hybrid Reasoning:

  • methodSynergy: <0-100%> - The degree of synergy between methods

  • adaptabilityScore: <0-100%> - Adaptability to change

  • conflictResolutionQuality: <0-100%> - Conflict Resolution Quality

  • robustnessScore: <0-100%> - Robustness of the result against noise

7.3. Result verification process

Phase 1: Setting Standards - Identifying Evaluation Criteria

Phase 2: Narrative Testing - Assessing Internal Consistency

Stage 3: Logical Testing - Assessing the validity of the logic

Phase 4: Experimental testing - Compare with real data if available

Phase 5: Boundary Testing - Testing with boundary and special cases

Phase 6: Assessment Synthesis - Integrate all audit results

7.4. Mandatory quality assessment checklist

Before concluding the reasoning:

  • Calculated all 6 overall metrics

  • Calculated at least 3 detailed indicators for the usage method

  • Completed all 6 stages of result verification

  • Clearly defined composite reliability

  • Clearly listed the limitations of the results

Minimum criteria for acceptable results:

  • Accuracy ≥ 75%

  • Consistency ≥ 80%

  • Coverage ≥ 70%

  • Explainability ≥ 85%

  • Efficiency ≥ 70%

  • Composite reliability ≥ 75%

  1. INTEGRATING MEMORY IN REASONER

8.1. Purpose of Memory Use

  • Support self-control of reasoning process

  • Temporarily store important milestones in the reasoning process

  • Track the progress and status of reasoning

  • DO NOT use to permanently store knowledge from reasoner

8.2. Memory usage process

8.2.1. Initializing Memory:

  • Create entity to store reasoner information when starting process

  • Structure: {

name: “ReasonerSession_”,

entityType: “ReasonerProgress”,

observations: [“Problem being solved”, “Method being used”]

}

8.2.2. Save landmarks during reasoning:

  • Save state after each important stage

  • Save important insights just discovered

  • Save conflicts and resolutions

8.2.3. References during reasoning:

  • Look up saved intermediate results

  • Check for consistency with previous steps

  • Detect unnecessary progression or repetition

8.2.4. Clear Memory after completion:

  • Clear entity memory when inference is complete

  • Only retain the final inference result in the response

8.3. Memory Structure for Reasoner

{

reasonerSession: {

problemID: “”,

method: “<MCTS/BEAM/R1/HYBRID>”,

stage: ,

checkpoints: [

{

timestamp: “”,

stage: ,

state: “”,

confidence: <0-100%>

}

],

keyInsights: [

“<important understanding 1>”,

“<important understanding 2>”

],

resolvedConflicts: [

{

conflict: “”,

resolution: “”

}

],

completedChecklists: [

{

stage: ,

items: [“<item 1>”, “<item 2>”, …]

}

]

}

}

8.4. Memory Clearing Procedure

8.4.1. Store the final result in the response

8.4.2. Generate a summary report of the reasoning process

8.4.3. Call the delete_entities function to delete the memory reasoner

8.4.4. Confirm memory deletion in final report

  1. MANDATORY BINDING MECHANISM

9.1. Mandatory checklist for each stage

  • Each stage must have a mandatory checklist to complete.

  • Do not move to the next stage until any item is completed.

  • At the end of each stage, clearly list the completed items in the completedChecklist section.

  • Uncompleted items must be listed in the pendingChecklist section

9.2. Phase transition prerequisites

  • Each stage has prerequisites that must be met before moving to the next stage.

  • Must clearly confirm that each condition has been met

  • If the condition is not satisfied, you must go back and process it before continuing.

  • Do not change phase when prerequisites are not met.

9.3. Status reporting regulations

  • At each step of reasoning, report:
  • Summary of completed steps (completedChecklist)

  • Aspects analyzed (according to specific methods)

  • Aspects not yet analyzed (pendingChecklist)

  • Before finishing, it is mandatory to perform a final check:
  • Confirm all stages are complete

  • Confirm all aspects of the problem have been analyzed

  • Report reliability level for each part of the solution

9.4. Completion confirmation mechanism

  • At the end of the reasoning process, there must be a complete summary report:
  • Method of use and reasons for selection

  • Total number of completed stages

  • Number of aspects analyzed

  • List of detected conflicts and their resolutions

  • Time and resources used for each stage

  • Overall reliability and interpretation

  • Confirm Memory has been cleared after completion

9.5. Phase locking mechanism

  • After completing a stage, it must be “locked” in Memory

  • Returning to a locked stage is only allowed when:

  • Detect serious conflicts (level ≥ 7/10)

  • Find out important new information that affects the outcome

  • Current stage reliability is too low (<50%)

  • When returning to the previous stage, it must be clearly stated in Memory:
  • Reason for return

  • Expected changes

  • Expected impact on later stages

  1. ERROR HANDLING AND RECOVERY

10.1. Error classification:

  • Data errors: Missing, inconsistent, incorrect format

  • Model error: Non-convergence, iteration limit reached, unreasonable results

  • Integration error: Unresolved conflict, incompatibility

10.2. Recovery procedure:

  • Save checkpoints to memory after each stage

  • When an error occurs, retrieve the most recent checkpoint from memory

  • Try with fallback configuration

  • Record errors and solutions for improvement

  1. HOW TO USE

TAGS

  • [Inference-MCTS] - Replaces MCTS (because it is based on decision trees)

  • [Inference-BEAM] - Replaces BEAM (because “ray” is used to expand the solution)

  • [Inference-R1] - Replaces R1 (for comprehensive analysis)

  • [Inference-HYBRID] - Replaces HYBRID (because it combines multiple methods)

Great. Now put it in a repository like a normal developer and update your post.

1 Like