KnowledgeRefinery/test_corpus/neural_networks.md
oho 38a99476d6 Knowledge Refinery: local-first semantic search & 3D concept visualization
macOS app for corpus ingestion, semantic search, and concept universe
visualization powered by local LLMs via LM Studio.

Architecture:
- Go daemon (17MB single binary, zero dependencies)
  - chi router, pure-Go SQLite, tiktoken tokenizer
  - 6-stage pipeline: scan → extract → chunk → embed → annotate → conceptualize
  - Brute-force cosine vector search in memory
  - 89 tests across 8 packages
- SwiftUI app (macOS 15+)
  - Multi-workspace management with auto-start daemons
  - Live pipeline progress, search, concept browser
  - WebGPU 3D universe renderer with Canvas2D fallback
  - Custom crystal app icon
2026-02-13 18:09:46 +01:00

2.9 KiB

Neural Networks: A Comprehensive Overview

History

The concept of artificial neural networks dates back to the 1940s when Warren McCulloch and Walter Pitts created a computational model for neural networks. In 1958, Frank Rosenblatt created the perceptron, an algorithm for pattern recognition based on a two-layer computer learning network.

The field experienced its first "AI winter" in the 1970s after Minsky and Papert published their book highlighting limitations of perceptrons. Interest revived in the 1980s with the development of backpropagation, and again in the 2010s with the advent of deep learning.

Architecture

A neural network consists of layers of interconnected nodes or neurons. The basic structure includes:

  • Input Layer: Receives the raw data
  • Hidden Layers: Process the data through weighted connections
  • Output Layer: Produces the final result

Each connection between neurons has a weight that is adjusted during training. The network learns by adjusting these weights to minimize the difference between predicted and actual outputs.

Activation Functions

Activation functions introduce non-linearity into the network, allowing it to learn complex patterns:

  • ReLU (Rectified Linear Unit): f(x) = max(0, x) - Most commonly used
  • Sigmoid: f(x) = 1/(1+e^(-x)) - Used for binary classification
  • Tanh: f(x) = (e^x - e^(-x))/(e^x + e^(-x)) - Centered around zero
  • Softmax: Used in output layer for multi-class classification

Training Process

  1. Forward Pass: Input data flows through the network to produce an output
  2. Loss Calculation: The difference between predicted and actual output is computed
  3. Backward Pass (Backpropagation): Gradients are computed for each weight
  4. Weight Update: Weights are adjusted using an optimizer (SGD, Adam, etc.)

Modern Architectures

Convolutional Neural Networks (CNNs)

Specialized for processing grid-like data (images). Use convolutional layers with filters that detect features like edges, textures, and shapes.

Recurrent Neural Networks (RNNs)

Designed for sequential data. Include memory that allows information to persist. LSTM and GRU variants address the vanishing gradient problem.

Transformers

Use self-attention mechanisms instead of recurrence. Enable parallel processing of sequences. Foundation for GPT, BERT, and other large language models.

Graph Neural Networks (GNNs)

Operate on graph-structured data. Useful for social networks, molecular structures, and knowledge graphs.

Applications

Neural networks are used in virtually every domain of AI:

  • Computer vision (image classification, object detection, segmentation)
  • Natural language processing (translation, summarization, question answering)
  • Speech recognition and synthesis
  • Drug discovery and molecular modeling
  • Autonomous vehicles
  • Financial forecasting
  • Game playing (AlphaGo, AlphaFold)