Neural Networks (COMP 258)

This course introduces artificial neural networks and deep learning models, emphasizing both theoretical foundations and practical implementation using modern machine learning frameworks.

Course Description

COMP 258 explores artificial neural networks and their application to real-world problems. Students study foundational neuron models, including McCulloch-Pitts, Hebbian learning, Perceptrons, ADALINE, and Multi-Layer Perceptrons (MLPs), before progressing to modern deep learning architectures.

Topics include convolutional neural networks (CNNs), sequence models such as recurrent neural networks (RNNs) and long short-term memory (LSTM) networks, attention mechanisms, and transformer-based models including Vision Transformers. Key learning components include backpropagation, activation functions, optimization methods, regularization techniques, and embedding layers.

Students gain hands-on experience using Keras and TensorFlow to design, train, and evaluate neural networks for tasks such as image recognition, natural language processing, and generative AI. AI-assisted development tools are introduced to support experimentation, debugging, and model refinement.

Course Learning Outcomes

Upon successful completion of this course, students will be able to:

Further Learning – Neural Networks and AI Frameworks

The following frameworks and research resources support continued learning and experimentation in neural networks and deep learning.

Course Materials

Primary Textbook

Géron, A. Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow (3rd ed.). O’Reilly Media, 2023.

Reference Textbooks

Mathematical and Supporting Resources

New Developments in Artificial Intelligence