Courses tagged with "Information environments" (59)
This course introduces the basic computational methods used to understand the cell on a molecular level. It covers subjects such as the sequence alignment algorithms: dynamic programming, hashing, suffix trees, and Gibbs sampling. Furthermore, it focuses on computational approaches to: genetic and physical mapping; genome sequencing, assembly, and annotation; RNA expression and secondary structure; protein structure and folding; and molecular interactions and dynamics.
This graduate level mathematics course covers decision theory, estimation, confidence intervals, and hypothesis testing. The course also introduces students to large sample theory. Other topics covered include asymptotic efficiency of estimates, exponential families, and sequential analysis.
This graduate-level course is an advanced introduction to applications and theory of numerical methods for solution of differential equations. In particular, the course focuses on physically-arising partial differential equations, with emphasis on the fundamental ideas underlying various methods.
This course provides an introduction to the theory and practice of quantum computation. Topics covered include: physics of information processing, quantum logic, quantum algorithms including Shor's factoring algorithm and Grover's search algorithm, quantum error correction, quantum communication, and cryptography.
This is an advanced topics course in model theory whose main theme is simple theories. We treat simple theories in the framework of compact abstract theories, which is more general than that of first order theories. We cover the basic properties of independence (i.e., non-dividing) in simple theories, the characterization of simple theories by the existence of a notion of independence, and hyperimaginary canonical bases.
This graduate level course is more extensive and theoretical treatment of the material in Computability, and Complexity (6.045J / 18.400J). Topics include Automata and Language Theory, Computability Theory, and Complexity Theory.
In this graduate-level course, we will be covering advanced topics in combinatorial optimization. We will start with non-bipartite matchings and cover many results extending the fundamental results of matchings, flows and matroids. The emphasis is on the derivation of purely combinatorial results, including min-max relations, and not so much on the corresponding algorithmic questions of how to find such objects. The intended audience consists of Ph.D. students interested in optimization, combinatorics, or combinatorial algorithms.
This course will give a detailed introduction to the theory of tensor categories and review some of its connections to other subjects (with a focus on representation-theoretic applications). In particular, we will discuss categorifications of such notions from ring theory as: module, morphism of modules, Morita equivalence of rings, commutative ring, the center of a ring, the centralizer of a subring, the double centralizer property, graded ring, etc.
The main goal of this course is to study the generalization ability of a number of popular machine learning algorithms such as boosting, support vector machines and neural networks. Topics include Vapnik-Chervonenkis theory, concentration inequalities in product spaces, and other elements of empirical process theory.
This course is organized around algorithmic issues that arise in machine learning. Modern machine learning systems are often built on top of algorithms that do not have provable guarantees, and it is the subject of debate when and why they work. In this class, we focus on designing algorithms whose performance we can rigorously analyze for fundamental machine learning problems.
Wavelets are localized basis functions, good for representing short-time events. The coefficients at each scale are filtered and subsampled to give coefficients at the next scale. This is Mallat's pyramid algorithm for multiresolution, connecting wavelets to filter banks. Wavelets and multiscale algorithms for compression and signal/image processing are developed. Subject is project-based for engineering and scientific applications.
This class covers the analysis and modeling of stochastic processes. Topics include measure theoretic probability, martingales, filtration, and stopping theorems, elements of large deviations theory, Brownian motion and reflected Brownian motion, stochastic integration and Ito calculus and functional limit theorems. In addition, the class will go over some applications to finance theory, insurance, queueing and inventory models.
Trusted paper writing service WriteMyPaper.Today will write the papers of any difficulty.