| 
                                              
                                                            
       
  | 
              
|                                                      
                                      
       [Announcements][Course Description][Class  Format][Grading][Contact
       Instructor/TA][Office Hours]
           | 
              
                             
     
      
  | 
         
|    This course is intended
  to  be  a broad exposure to the field of machine learning - a topic which,
  in  essence,   deals with the issue of designing machines,  algorithms
and   tools  which automatically  improve with experience.  This course 
is designed   to  give a graduate-level  student a thorough grounding in
the methodologies,          technologies,        mathematics
and algorithms     currently needed by people  who do research in
learning with data. The  course   will initially discuss  the fundamental
issues in learning theory  including   topics like optimization,  cost function,
information gain, regularization,     model selection, VC dimension  and
PAC learnability. Topics from Inductive     and Analytical learning theory
 including Decision trees, Boosting &    Bagging,   CART and concept
learning/hypotheses  evaluation will be addressed.    Then, we will look
at various forms of supervised  learning implementations    - both linear
(Projection methods, LDA, subspace)  and non-linear methods    (RBFs, Neural
Networks, SVMs, LWL) while also addressing  issues of subset/feature    
selection and dimensionality reduction. Topics from unspervised learning
   like ICA, Factor Analysis and Clustering will be covered in addition to
 basics  of Reinforcement Learning theory .                             
                                                  
       Skills from this course will be useful for basic and applied research in the field of Computer Science(e.g. algorithms, data mining), Statistics(e.g. optimization theory, regression), Artificial Intelligence(e.g. pattern recognition, data visualization), Information theory, Finance(e.g. time series prediction) and Cognitive Science.  | 
               
|                             
      The  syllabus is aimed at covering the state  of the art developments 
 in  the machine  learning community. This will be achieved  by first identifying 
   and asking  the 'right questions' about fundamental  issues that 
 are  critical in determining the success or failure of a learning  system. 
 Then,        implementations   under different domains, requirements 
  and  constraints will be explored. | 
               |
| Fundamental 
    Issues in Learning   Theory | 
             |
| Optimization & 
 Cost   function       -   Least Squares, Minimum-Variance Minimum-Bias
   Estimates,        Lagrange Methods  | 
             |
| Information Gain 
&    Function  Spaces  | 
             |
| VC Dimension & 
 PAC   learnability | 
             |
| Regularization  - Green 
    function, P-norm  regularization,  Smootheness criterion,  Bias Variance Tradeoff  | 
             |
| Model Selection - MDL, Bayesian Information Criterion, AIC, Cross 
      Validation | 
             |
| Inductive 
    and Analytical Learning  | 
             |
| Decision Trees- 
    MARS, Decision  tree representation, overfitting | 
             |
| Hypotheses Evaluation | 
             |
| Learning Rule Sets- 
    FOIL | 
             |
| Explanation Based
 Learning    (EBL)-   Deductive learning, Inductive Bias and
 Knowledge  Level         learning  | 
             |
| Bootstrap methods- 
    Boosting, Bagging  etc.  | 
             |
| Supervised 
    Learning -   Regression and Classification | 
             |
| Linear Methods | 
             |
| Non-linear Methods 
           | 
             |
| Feature Selection
 &    Dimensionality Reduction-   Global vs Local
 and Projection    methods,        Automatic Relevance Detection (ARD)  | 
             |
| Genetic Algorithms 
       - Fitness function, Genetic Programming 
        | 
             |
| Unsupervised 
    Learning Clustering,   Vector Quantization, SOMs, 
 Principa   Curves, ICA,  Projection Pursuit, Factor Analysis and Latent Variables  | 
             |
| Reinforcement 
    Learning - Dynamic Programming,
    Monte Carlo   & TD learning     | 
             |
| Instructor | 
          Teaching Assistant | 
          Grader | 
        
| Prof. Sethu Vijayakumar
        Research Assistant Professor University of Southern California Hedco Neurosciences Building HNB-103 Los Angeles, CA 90089-2520 phone: (213) 740 1551 email: sethu@usc.edu  | 
                    Aaron D'Souza  Graduate Research Assistant University of Southern California Hedco Neurosciences Building HNB-001 Los Angeles, CA 90089-2520 phone: (213) 821 6370 email: adsouza@usc.edu  | 
                    Rohan Chitradurga Masters Course Student (EECN), Department of Electrical Engineering, University of Southern California Hedco Neurosciences Building HNB-001 Los Angeles, CA 90089-2520 email: chitradu@usc.edu  | 
        
                                      
                                      
      Class FormatThe course consists of lectures 
     with discussions, reading assignments, and homework assignments. In addition,
     there will be time allocated for review of some of the latest seminal
 developments    ( e.g. papers published at ICML, NIPS) in key areas. There
 will be a mid-term exam reviewing the basics of the course. At the end of
 the course, each student will carry   out an independent final project in
 an area subject to the instructor's  approval.   
                                                                        
                           
      Grading
 PrerequisitesIdeally CS561 or CS442, basic knowledge in linear algebra, statistics, calculus, and programming in MATLAB/C (or another language), or permission by instructor. Working knowledge of probability theory and algorithms is recommended but not required.Academic IntegrityAll students are required 
     to abide by the USC
       code of Academic Integrity. Violation of that Code will be dealt
  with     as described in SCAMPUS. If you have any questions about the responsibilities
       of either students, faculty, or graders under this policy, contact
the          instructor or the
      Office
 of      Student Conduct. 
                                                                        
                  
      Disabilities and Academic AccomodationStudents requesting academic 
     accomodations based on a disability are required to register with Disability
Services and Programs (DSP) each semester. A letter of verification for
approved accomodations can be obtained from DSP when adequate documentaion
is filed. Please be sure the letter is delivered to the instructor (or TA)
as early in the semester as possible. DSP is open Monday-Friday, 8:30-5:00.
The office is in Student Union 301 and their phone number is (213) 740-0776. 
                  | 
               
              
             If you have comments or suggestions, send email to sethu@usc.edu