Safetensors
llama
llama3
context-8000
layer-fusion-conceptual
tensor-fusion-conceptual
bias-removal
decode
coherence-enhancement
custom-code
grouping
reward-alignment
reasoning-tuned
tool-use-hint
long-context-hint
memory-hint
conceptual-graph-hint
emotional-intelligence-hint
ethical-alignment-hint
causal-inference-hint
planning-hint
situational-awareness-hint
creativity-hint
learning-adaptivity-hint
knowledge-graph-hint
theory-of-mind-hint
self-correction-hint
uncertainty-quantification-hint
interpretability-hint
bias-mitigation-hint
context-compression-hint
abstraction-control-hint
novelty-detection-hint
explainability-hint
instruct
adaptive-memory-hint
goal-driven-hint
hierarchical-reasoning-hint
symbolic-representation-hint
embodied-simulation-hint
ethical-reasoning-hint
proactive-behavior-hint
explainability-levels-hint
rl-integration-hint
fl-compatibility-hint
dp-features-hint
robustness-hint
calibration-hint
ood-detection-hint
custom_code
Add modeling_custom.py
Browse files- modeling_custom.py +6 -1
modeling_custom.py
CHANGED
|
@@ -8,5 +8,10 @@ logger = logging.get_logger(__name__)
|
|
| 8 |
class CustomLlamaForCausalLM(LlamaForCausalLM):
|
| 9 |
def __init__(self, config: LlamaConfig):
|
| 10 |
super().__init__(config)
|
| 11 |
-
logger.info("CustomLlamaForCausalLM initialized.")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 12 |
|
|
|
|
| 8 |
class CustomLlamaForCausalLM(LlamaForCausalLM):
|
| 9 |
def __init__(self, config: LlamaConfig):
|
| 10 |
super().__init__(config)
|
| 11 |
+
logger.info("CustomLlamaForCausalLM initialized with conceptual features documented in config.")
|
| 12 |
+
if getattr(config, 'conceptual_features', {}).get('grouping_logic'):
|
| 13 |
+
logger.info(f"Conceptual grouping logic enabled with size {config.conceptual_features.get('group_size', 'N/A')}")
|
| 14 |
+
if getattr(config, 'conceptual_features', {}).get('long_context_optimization'):
|
| 15 |
+
logger.info("Conceptual long context optimization hint detected.")
|
| 16 |
+
|
| 17 |
|