Research Article | | Peer-Reviewed

Algebraic Cekirge Method for Deterministic and Energy-efficient Transformer Language Models

Received: 25 October 2025     Accepted: 5 November 2025     Published: 22 November 2025
Views:       Downloads:
Abstract

This paper introduces a unified deterministic algebraic framework for transformer-style language modeling, extending the σ-Based (Cekirge) methodology toward energy-efficient and interpretable computation. The approach constructs σ-regularized, nonsingular matrices for queries, keys, values, and output weights (Q, K, V, W), enabling attention and decoding to be computed in closed form without iterative stochastic gradient descent. By enforcing σ-regularization, matrix invertibility and numerical stability are guaranteed, allowing direct algebraic determination of weights from paired input–output examples rather than optimization through back propagation. Analytical and numerical experiments, including a controlled five-token model, demonstrate that the deterministic algebraic solution reproduces the predictive behavior of gradient descent while eliminating randomness and reducing computation time by more than sixty-fold. The framework unifies several complementary formulations—the Four-Cluster Deterministic Map, Frozen-Library Forward Training, σ-Matrix Fast Learning, and Inverse Deterministic Energy-Saving Training (IDEST)—each contributing to a closed, energy-saving learning process that transforms optimization into deterministic algebraic resolution. Conceptually, the Cekirge Method parallels perceptual refinement: just as John remains himself while seen through blue eyeglasses but becomes obscured behind a wooden mask. The model’s internal mappings preserve identity under small deterministic perturbations (ε) but lose interpretability under stochastic noise. This analogy captures the method’s central philosophy—learning through controlled perturbation that reveals structure without destroying equilibrium. Finally, the study situates deterministic computation within a sustainability perspective. Global energy consumption, intensified by iterative AI training, has surpassed ecological thresholds. The Cekirge framework reconceives learning as a finite algebraic equilibrium rather than an energy-intensive iterative loop, aligning computational intelligence with thermodynamic and social efficiency. It proposes that future AI systems should pursue mathematical determinism and ecological responsibility in parallel, ensuring progress that is both computationally exact and energetically sustainable. A supplementary real-data experiment confirms that deterministic algebraic decoding achieves comparable accuracy to gradient descent while operating with markedly lower computational energy.

Published in American Journal of Artificial Intelligence (Volume 9, Issue 2)
DOI 10.11648/j.ajai.20250902.25
Page(s) 258-271
Creative Commons

This is an Open Access article, distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution and reproduction in any medium or format, provided the original work is properly cited.

Copyright

Copyright © The Author(s), 2025. Published by Science Publishing Group

Keywords

Deterministic Learning, σ-regularization, Algebraic AI, Energy-efficient Computation, Transformer Models, Cekirge Method, Closed-form Learning, Sustainable Machine Intelligence

References
[1] Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, Ł., and Polosukhin, I. “Attention Is All You Need,” NeurIPS, 2017.
[2] Goodfellow, I., Bengio, Y., and Courville, A. “Deep Learning,” MIT Press, 2016.
[3] Kingma, D., and Ba, J. “Adam Optimizer,” ICLR, 2015.
[4] Schmidhuber, J. “Deep Learning in Neural Networks: An Overview,” Neural Networks, 61, 85–117, 2015.
[5] Devlin, J., Chang, M.-W., Lee, K., and Toutanova, K. “BERT: Pre-training of Deep Bidirectional Transformers,” NAACL-HLT, 2019.
[6] Bengio, Y. “Learning Deep Architectures for AI,” Foundations and Trends in Machine Learning, 2(1), 1–127, 2009.
[7] Hinton, G. E., “Deep Learning and Representations,” Nature, 521(7553), 436–444, 2015.
[8] LeCun, Y., Bottou, L., Orr, G., and Müller, K. “Efficient Backpropagation,” Springer, 1998.
[9] Cekirge, H. M. “An Alternative Way of Determining Biases and Weights for the Training of Neural Networks,” American Journal of Artificial Intelligence, 9(2), 129–132, 2025.
[10] Cekirge, H. M. “Algebraic σ-Based (Cekirge) Model for Deterministic and Efficient Unsupervised Machine Learning,” American Journal of Artificial Intelligence, 9(2), 198–205, 2025.
[11] Cekirge, H. M. “Cekirge’s σ-Based ANN Model for Deterministic, Energy-Efficient, Scalable AI with Large-Matrix Capability,” American Journal of Artificial Intelligence, 9(2), 206–216, 2025.
[12] Katharopoulos, A., Vyas, A., Pappas, N., and Fleuret, F. “Transformers are RNNs: Fast autoregressive Transformers with linear attention,” ICML. PMLR, 5156–5165, 2020.
[13] Schlag, I., Irie, K., and Schmidhuber, J. “Linear Transformers Are Secretly Fast Weight Programmers,” ICML. Springer, 9355–9366, 2021.
[14] Mikolov, T., Chen, K., Corrado, G., and Dean, J. “Efficient estimation of word representations in vector space,” 2013.
[15] Glorot, X., and Bengio, Y. “Understanding the difficulty of training deep feedforward neural networks,” Journal of Machine Learning Research, 9, 249–256, 2010.
[16] Goldberg, Y. “A Primer on Neural Network Models for Natural Language Processing,” Journal of Artificial Intelligence Research, 57, 345–420, 2016.
[17] Shaw, P., Uszkoreit, J., and Vaswani, A. “Self-Attention with Relative Position Representations,” 2018. arXiv: 1803.02155.
[18] Friston, K., A free energy principle for the brain. Journal of Physiology-Paris, 100(1–3), 70–87, 2006.
[19] Schmidhuber, J., Formal theory of creativity, fun, and intrinsic motivation (1990–2010). IEEE Transactions on Autonomous Mental Development, 2(3), 230–247, 2010.
[20] IEA. “Tracking Clean Energy Progress 2022,” Paris: OECD/IEA, 2022.
[21] Intergovernmental Panel on Climate Change (IPCC), Climate Change 2023: Synthesis Report. Geneva: IPCC, 2023.
[22] International Energy Agency (IEA). (2024). World Energy Outlook 2024. Paris: OECD/IEA, 2024.
[23] Rockström, J., Gupta, J., and Dubash, N. K., The world’s biggest challenge: Energy transition under planetary boundaries. Science, 370(6521), 36–40, 2020.
[24] Sovacool, B. K., Hess, D. J., & Baldwin, E., Socio-technical transitions in energy. Annual Review of Environment and Resources, 46, 209–236, 2021.
[25] IEA, “Renewables 2023: Analysis and Forecast to 2028.” Paris: OECD/IEA, 2023
[26] World Bank. “Global Energy Data and Sustainable Development Indicators,”Washington, D.C. (2023).
[27] Markard, J. “The energy transition as a socio-technical process,” Energy Research and Social Science, 98, 102978, 2023.
[28] Jevons, W. S. “The Coal Question: An Inquiry Concerning the Progress of the Nation,” London: Macmillan, 1865.
[29] Santarius, T., Walnum, H. J., and Aall, C. “Rethinking climate rebound,” Springer, 2018.
[30] United Nations. “Sustainable Development Goals Report 2023,” New York: United Nations, 2023.
Cite This Article
  • APA Style

    Cekirge, H. M. (2025). Algebraic Cekirge Method for Deterministic and Energy-efficient Transformer Language Models. American Journal of Artificial Intelligence, 9(2), 258-271. https://doi.org/10.11648/j.ajai.20250902.25

    Copy | Download

    ACS Style

    Cekirge, H. M. Algebraic Cekirge Method for Deterministic and Energy-efficient Transformer Language Models. Am. J. Artif. Intell. 2025, 9(2), 258-271. doi: 10.11648/j.ajai.20250902.25

    Copy | Download

    AMA Style

    Cekirge HM. Algebraic Cekirge Method for Deterministic and Energy-efficient Transformer Language Models. Am J Artif Intell. 2025;9(2):258-271. doi: 10.11648/j.ajai.20250902.25

    Copy | Download

  • @article{10.11648/j.ajai.20250902.25,
      author = {Huseyin Murat Cekirge},
      title = {Algebraic Cekirge Method for Deterministic and Energy-efficient Transformer Language Models
    },
      journal = {American Journal of Artificial Intelligence},
      volume = {9},
      number = {2},
      pages = {258-271},
      doi = {10.11648/j.ajai.20250902.25},
      url = {https://doi.org/10.11648/j.ajai.20250902.25},
      eprint = {https://article.sciencepublishinggroup.com/pdf/10.11648.j.ajai.20250902.25},
      abstract = {This paper introduces a unified deterministic algebraic framework for transformer-style language modeling, extending the σ-Based (Cekirge) methodology toward energy-efficient and interpretable computation. The approach constructs σ-regularized, nonsingular matrices for queries, keys, values, and output weights (Q, K, V, W), enabling attention and decoding to be computed in closed form without iterative stochastic gradient descent. By enforcing σ-regularization, matrix invertibility and numerical stability are guaranteed, allowing direct algebraic determination of weights from paired input–output examples rather than optimization through back propagation. Analytical and numerical experiments, including a controlled five-token model, demonstrate that the deterministic algebraic solution reproduces the predictive behavior of gradient descent while eliminating randomness and reducing computation time by more than sixty-fold. The framework unifies several complementary formulations—the Four-Cluster Deterministic Map, Frozen-Library Forward Training, σ-Matrix Fast Learning, and Inverse Deterministic Energy-Saving Training (IDEST)—each contributing to a closed, energy-saving learning process that transforms optimization into deterministic algebraic resolution. Conceptually, the Cekirge Method parallels perceptual refinement: just as John remains himself while seen through blue eyeglasses but becomes obscured behind a wooden mask. The model’s internal mappings preserve identity under small deterministic perturbations (ε) but lose interpretability under stochastic noise. This analogy captures the method’s central philosophy—learning through controlled perturbation that reveals structure without destroying equilibrium. Finally, the study situates deterministic computation within a sustainability perspective. Global energy consumption, intensified by iterative AI training, has surpassed ecological thresholds. The Cekirge framework reconceives learning as a finite algebraic equilibrium rather than an energy-intensive iterative loop, aligning computational intelligence with thermodynamic and social efficiency. It proposes that future AI systems should pursue mathematical determinism and ecological responsibility in parallel, ensuring progress that is both computationally exact and energetically sustainable. A supplementary real-data experiment confirms that deterministic algebraic decoding achieves comparable accuracy to gradient descent while operating with markedly lower computational energy.
    },
     year = {2025}
    }
    

    Copy | Download

  • TY  - JOUR
    T1  - Algebraic Cekirge Method for Deterministic and Energy-efficient Transformer Language Models
    
    AU  - Huseyin Murat Cekirge
    Y1  - 2025/11/22
    PY  - 2025
    N1  - https://doi.org/10.11648/j.ajai.20250902.25
    DO  - 10.11648/j.ajai.20250902.25
    T2  - American Journal of Artificial Intelligence
    JF  - American Journal of Artificial Intelligence
    JO  - American Journal of Artificial Intelligence
    SP  - 258
    EP  - 271
    PB  - Science Publishing Group
    SN  - 2639-9733
    UR  - https://doi.org/10.11648/j.ajai.20250902.25
    AB  - This paper introduces a unified deterministic algebraic framework for transformer-style language modeling, extending the σ-Based (Cekirge) methodology toward energy-efficient and interpretable computation. The approach constructs σ-regularized, nonsingular matrices for queries, keys, values, and output weights (Q, K, V, W), enabling attention and decoding to be computed in closed form without iterative stochastic gradient descent. By enforcing σ-regularization, matrix invertibility and numerical stability are guaranteed, allowing direct algebraic determination of weights from paired input–output examples rather than optimization through back propagation. Analytical and numerical experiments, including a controlled five-token model, demonstrate that the deterministic algebraic solution reproduces the predictive behavior of gradient descent while eliminating randomness and reducing computation time by more than sixty-fold. The framework unifies several complementary formulations—the Four-Cluster Deterministic Map, Frozen-Library Forward Training, σ-Matrix Fast Learning, and Inverse Deterministic Energy-Saving Training (IDEST)—each contributing to a closed, energy-saving learning process that transforms optimization into deterministic algebraic resolution. Conceptually, the Cekirge Method parallels perceptual refinement: just as John remains himself while seen through blue eyeglasses but becomes obscured behind a wooden mask. The model’s internal mappings preserve identity under small deterministic perturbations (ε) but lose interpretability under stochastic noise. This analogy captures the method’s central philosophy—learning through controlled perturbation that reveals structure without destroying equilibrium. Finally, the study situates deterministic computation within a sustainability perspective. Global energy consumption, intensified by iterative AI training, has surpassed ecological thresholds. The Cekirge framework reconceives learning as a finite algebraic equilibrium rather than an energy-intensive iterative loop, aligning computational intelligence with thermodynamic and social efficiency. It proposes that future AI systems should pursue mathematical determinism and ecological responsibility in parallel, ensuring progress that is both computationally exact and energetically sustainable. A supplementary real-data experiment confirms that deterministic algebraic decoding achieves comparable accuracy to gradient descent while operating with markedly lower computational energy.
    
    VL  - 9
    IS  - 2
    ER  - 

    Copy | Download

Author Information
  • Sections