Cardiac conduction systemThe cardiac conduction system (CCS) (also called the electrical conduction system of the heart) transmits the signals generated by the sinoatrial node – the heart's pacemaker, to cause the heart muscle to contract, and pump blood through the body's circulatory system. The pacemaking signal travels through the right atrium to the atrioventricular node, along the bundle of His, and through the bundle branches to Purkinje fibers in the walls of the ventricles. The Purkinje fibers transmit the signals more rapidly to stimulate contraction of the ventricles.
Residual neural networkA Residual Neural Network (a.k.a. Residual Network, ResNet) is a deep learning model in which the weight layers learn residual functions with reference to the layer inputs. A Residual Network is a network with skip connections that perform identity mappings, merged with the layer outputs by addition. It behaves like a Highway Network whose gates are opened through strongly positive bias weights. This enables deep learning models with tens or hundreds of layers to train easily and approach better accuracy when going deeper.
Convolutional neural networkConvolutional neural network (CNN) is a regularized type of feed-forward neural network that learns feature engineering by itself via filters (or kernel) optimization. Vanishing gradients and exploding gradients, seen during backpropagation in earlier neural networks, are prevented by using regularized weights over fewer connections. For example, for each neuron in the fully-connected layer 10,000 weights would be required for processing an image sized 100 × 100 pixels.
Electrophysiology studyA cardiac electrophysiology study (EP test or EP study) is a minimally invasive procedure using catheters introduced through a vein or artery to record electrical activity from within the heart. This electrical activity is recorded when the heart is in a normal rhythm (sinus rhythm) to assess the conduction system of the heart and to look for additional electrical connections (accessory pathways), and during any abnormal heart rhythms that can be induced.
Cardiac muscleCardiac muscle (also called heart muscle or myocardium) is one of three types of vertebrate muscle tissues, with the other two being skeletal muscle and smooth muscle. It is an involuntary, striated muscle that constitutes the main tissue of the wall of the heart. The cardiac muscle (myocardium) forms a thick middle layer between the outer layer of the heart wall (the pericardium) and the inner layer (the endocardium), with blood supplied via the coronary circulation.
Machine learningMachine learning (ML) is an umbrella term for solving problems for which development of algorithms by human programmers would be cost-prohibitive, and instead the problems are solved by helping machines 'discover' their 'own' algorithms, without needing to be explicitly told what to do by any human-developed algorithms. Recently, generative artificial neural networks have been able to surpass results of many previous approaches.
Cardiac pacemakerThe contraction of cardiac muscle (heart muscle) in all animals is initiated by electrical impulses known as action potentials that in the heart are known as cardiac action potentials. The rate at which these impulses fire controls the rate of cardiac contraction, that is, the heart rate. The cells that create these rhythmic impulses, setting the pace for blood pumping, are called pacemaker cells, and they directly control the heart rate. They make up the cardiac pacemaker, that is, the natural pacemaker of the heart.
Membrane potentialMembrane potential (also transmembrane potential or membrane voltage) is the difference in electric potential between the interior and the exterior of a biological cell. That is, there is a difference in the energy required for electric charges to move from the internal to exterior cellular environments and vice versa, as long as there is no acquisition of kinetic energy or the production of radiation. The concentration gradients of the charges directly determine this energy requirement.
Transformer (machine learning model)A transformer is a deep learning architecture that relies on the parallel multi-head attention mechanism. The modern transformer was proposed in the 2017 paper titled 'Attention Is All You Need' by Ashish Vaswani et al., Google Brain team. It is notable for requiring less training time than previous recurrent neural architectures, such as long short-term memory (LSTM), and its later variation has been prevalently adopted for training large language models on large (language) datasets, such as the Wikipedia corpus and Common Crawl, by virtue of the parallelized processing of input sequence.
Feature learningIn machine learning, feature learning or representation learning is a set of techniques that allows a system to automatically discover the representations needed for feature detection or classification from raw data. This replaces manual feature engineering and allows a machine to both learn the features and use them to perform a specific task. Feature learning is motivated by the fact that machine learning tasks such as classification often require input that is mathematically and computationally convenient to process.