Meta-Learning and Universality

Σχετικά έγγραφα
2 Composition. Invertible Mappings

C.S. 430 Assignment 6, Sample Solutions

derivation of the Laplacian from rectangular to spherical coordinates

Other Test Constructions: Likelihood Ratio & Bayes Tests

ST5224: Advanced Statistical Theory II

EE512: Error Control Coding

Lecture 2: Dirac notation and a review of linear algebra Read Sakurai chapter 1, Baym chatper 3

3.4 SUM AND DIFFERENCE FORMULAS. NOTE: cos(α+β) cos α + cos β cos(α-β) cos α -cos β

SCHOOL OF MATHEMATICAL SCIENCES G11LMA Linear Mathematics Examination Solutions

Matrices and Determinants

Fractional Colorings and Zykov Products of graphs

Partial Differential Equations in Biology The boundary element method. March 26, 2013

ΚΥΠΡΙΑΚΗ ΕΤΑΙΡΕΙΑ ΠΛΗΡΟΦΟΡΙΚΗΣ CYPRUS COMPUTER SOCIETY ΠΑΓΚΥΠΡΙΟΣ ΜΑΘΗΤΙΚΟΣ ΔΙΑΓΩΝΙΣΜΟΣ ΠΛΗΡΟΦΟΡΙΚΗΣ 19/5/2007

w o = R 1 p. (1) R = p =. = 1

6.3 Forecasting ARMA processes

Problem Set 3: Solutions

HOMEWORK 4 = G. In order to plot the stress versus the stretch we define a normalized stretch:

Finite Field Problems: Solutions

Section 8.3 Trigonometric Equations

CHAPTER 25 SOLVING EQUATIONS BY ITERATIVE METHODS

Solution Series 9. i=1 x i and i=1 x i.

Homework 8 Model Solution Section

Approximation of distance between locations on earth given by latitude and longitude

Phys460.nb Solution for the t-dependent Schrodinger s equation How did we find the solution? (not required)

Lecture 2. Soundness and completeness of propositional logic

Section 7.6 Double and Half Angle Formulas

Congruence Classes of Invertible Matrices of Order 3 over F 2

Nowhere-zero flows Let be a digraph, Abelian group. A Γ-circulation in is a mapping : such that, where, and : tail in X, head in

Takeaki Yamazaki (Toyo Univ.) 山崎丈明 ( 東洋大学 ) Oct. 24, RIMS

Math221: HW# 1 solutions

Statistical Inference I Locally most powerful tests

Example Sheet 3 Solutions

The Simply Typed Lambda Calculus

Bayesian statistics. DS GA 1002 Probability and Statistics for Data Science.

Numerical Analysis FMN011

4.6 Autoregressive Moving Average Model ARMA(1,1)

Jesse Maassen and Mark Lundstrom Purdue University November 25, 2013

Απόκριση σε Μοναδιαία Ωστική Δύναμη (Unit Impulse) Απόκριση σε Δυνάμεις Αυθαίρετα Μεταβαλλόμενες με το Χρόνο. Απόστολος Σ.

Tridiagonal matrices. Gérard MEURANT. October, 2008

Srednicki Chapter 55

Solutions to Exercise Sheet 5

Instruction Execution Times

Areas and Lengths in Polar Coordinates

Exercises 10. Find a fundamental matrix of the given system of equations. Also find the fundamental matrix Φ(t) satisfying Φ(0) = I. 1.

What happens when two or more waves overlap in a certain region of space at the same time?

Εργαστήριο Ανάπτυξης Εφαρμογών Βάσεων Δεδομένων. Εξάμηνο 7 ο

ΚΥΠΡΙΑΚΗ ΕΤΑΙΡΕΙΑ ΠΛΗΡΟΦΟΡΙΚΗΣ CYPRUS COMPUTER SOCIETY ΠΑΓΚΥΠΡΙΟΣ ΜΑΘΗΤΙΚΟΣ ΔΙΑΓΩΝΙΣΜΟΣ ΠΛΗΡΟΦΟΡΙΚΗΣ 6/5/2006

Econ 2110: Fall 2008 Suggested Solutions to Problem Set 8 questions or comments to Dan Fetter 1

ΠΤΥΧΙΑΚΗ ΕΡΓΑΣΙΑ ΒΑΛΕΝΤΙΝΑ ΠΑΠΑΔΟΠΟΥΛΟΥ Α.Μ.: 09/061. Υπεύθυνος Καθηγητής: Σάββας Μακρίδης

Overview. Transition Semantics. Configurations and the transition relation. Executions and computation

k A = [k, k]( )[a 1, a 2 ] = [ka 1,ka 2 ] 4For the division of two intervals of confidence in R +

6.1. Dirac Equation. Hamiltonian. Dirac Eq.

Ordinal Arithmetic: Addition, Multiplication, Exponentiation and Limit

EE101: Resonance in RLC circuits

Homework 3 Solutions

Math 6 SL Probability Distributions Practice Test Mark Scheme

Space-Time Symmetries

[1] P Q. Fig. 3.1

Reminders: linear functions

Example of the Baum-Welch Algorithm

Chapter 6: Systems of Linear Differential. be continuous functions on the interval

( ) 2 and compare to M.

Concrete Mathematics Exercises from 30 September 2016

Access Control Encryption Enforcing Information Flow with Cryptography

These derivations are not part of the official forthcoming version of Vasilaky and Leonard

Areas and Lengths in Polar Coordinates

1. For each of the following power series, find the interval of convergence and the radius of convergence:

Section 9.2 Polar Equations and Graphs

Exercises to Statistics of Material Fatigue No. 5

Practice Exam 2. Conceptual Questions. 1. State a Basic identity and then verify it. (a) Identity: Solution: One identity is csc(θ) = 1

CHAPTER 101 FOURIER SERIES FOR PERIODIC FUNCTIONS OF PERIOD

New bounds for spherical two-distance sets and equiangular lines

The challenges of non-stable predicates

Second Order Partial Differential Equations

Pg The perimeter is P = 3x The area of a triangle is. where b is the base, h is the height. In our case b = x, then the area is

Mean bond enthalpy Standard enthalpy of formation Bond N H N N N N H O O O

EE434 ASIC & Digital Systems Arithmetic Circuits

Fourier Series. MATH 211, Calculus II. J. Robert Buchanan. Spring Department of Mathematics

ANSWERSHEET (TOPIC = DIFFERENTIAL CALCULUS) COLLECTION #2. h 0 h h 0 h h 0 ( ) g k = g 0 + g 1 + g g 2009 =?

MATH423 String Theory Solutions 4. = 0 τ = f(s). (1) dτ ds = dxµ dτ f (s) (2) dτ 2 [f (s)] 2 + dxµ. dτ f (s) (3)

forms This gives Remark 1. How to remember the above formulas: Substituting these into the equation we obtain with

Electronic Analysis of CMOS Logic Gates

Computing Gradient. Hung-yi Lee 李宏毅

PARTIAL NOTES for 6.1 Trigonometric Identities

Inverse trigonometric functions & General Solution of Trigonometric Equations

If we restrict the domain of y = sin x to [ π, π ], the restrict function. y = sin x, π 2 x π 2

Living and Nonliving Created by: Maria Okraska

Assalamu `alaikum wr. wb.

Lecture 10 - Representation Theory III: Theory of Weights

TMA4115 Matematikk 3

If we restrict the domain of y = sin x to [ π 2, π 2

Lecture 13 - Root Space Decomposition II

2. THEORY OF EQUATIONS. PREVIOUS EAMCET Bits.

D Alembert s Solution to the Wave Equation

Συστήματα Διαχείρισης Βάσεων Δεδομένων

Μηχανική Μάθηση Hypothesis Testing

Capacitors - Capacitance, Charge and Potential Difference

Εγκατάσταση λογισμικού και αναβάθμιση συσκευής Device software installation and software upgrade

b. Use the parametrization from (a) to compute the area of S a as S a ds. Be sure to substitute for ds!

Transcript:

Meta-Learning and Universality Conference paper at ICLR 2018 Chelsea Finn & Sergey Levine (US Berkeley) Youngseok Yoon

Contents The Universality in Meta-Learning Model Construct (Pre-update) Single gradient step and Post-update Show universality from model Experiments Conclusion

The Universality in Meta-Learning What is the universality in Meta-Learning? Universality in Neural Network: Approximate any function ff: xx yy Universality in Meta-Learning: Approximate any function ff: xx, yy kk, xx yy Input: The training data xx, yy kk, and Test data xx. In this presentation, only show about kk = 1 case. Our model to verify Universality: MAML: ff ; θθ ff( ; θθ ), θθ = θθ αα θθ L tttttttttttt, ff ; θθ Update model ff( ; θθ) using gradient update

Model Construct (Pre-update) Our model: Deep neural network with ReLU, MAML training.

Model Construct (Pre-update) Construct model ff ; θθ with ReLU If we assume inputs and all pre-synaptic activations are nonnegative, then deep ReLU act like deep linear networks. (1) ff ; θθ = ff oooooo NN ii=1 WW ii φφ ; θθ ffff, θθ bb ; θθ oooooo

Single gradient step and Post-update WWii L = aa ii bb ii 1 aa ii : the error gradient with respect to the pre-synaptic activations at layer ii bb ii 1 : the forward post-synaptic activations at layer ii 1 ffffffffffffff WW ii+1 bb ii 1 WW ii aa ii WW ii 1

Single gradient step and Post-update let zz = Then, NN ii=1 WW ii φφ xx; θθ ffff, θθ bb, zz L = ee(xx, yy) WWii L = aa ii bb ii 1 = WW ii 1 WW NN 1 ee xx, yy jj=ii+1 WW jj φφ xx; θθ ffff, θθ bb = ii 1 jj=1 WW jj ee xx, yy φφ xx; NN θθffff, θθ bb jj=ii+1 WW jj ffffffffffffff WW ii+1 bb ii 1 WW ii aa ii WW ii 1

Single gradient step and Post-update let zz = Then, NN ii=1 WW ii φφ xx; θθ ffff, θθ bb, zz L = ee(xx, yy) WWii L = aa ii bb ii 1 = WW ii 1 WW NN 1 ee xx, yy jj=ii+1 WW jj φφ xx; θθ ffff, θθ bb = ii 1 jj=1 WW jj ee xx, yy φφ xx; NN θθffff, θθ bb jj=ii+1 WW jj Therefore, post-update value of NN ii=1 NN ii=1 NN WW ii αασ ii=1 ii 1 jj=1 WW jj WW NN ii = ii=1 ii 1 WW jj ee xx, yy φφ xx; θθffff, θθ bb jj=ii+1 NN jj=1 WW ii αα WWii L WW ii 1 jj jj=1 WW jj OO αα 2

Single gradient step and Post-update Post-update value of NN ii=1 NN ii=1 NN WW ii αασ ii=1 ii 1 jj=1 WW jj jj=1 WW NN ii = ii=1 WW ii αα WWii L ii 1 WW jj ee xx, yy φφ xx; θθffff, θθ bb jj=ii+1 NN WW ii 1 jj jj=1 WW jj OO αα 2 Now, post-update value of zz when xx input into ff ; θθ NN zz = ii=1 WW ii φφ xx ; θθ ffff, θθ bb = NN ii=1 WW ii φφ xx ; θθ ffff, θθ bb NN ii 1 WW jj ee xx, yy φφ xx; NN θθffff, θθ bb jj=ii+1 αασ ii=1 ii 1 jj=1 WW jj And, ff xx ; θθ jj=1 = ff oooooo zz ; θθ oooooo WW ii 1 jj jj=1 WW jj φφ xx ; θθ ffff, θθ bb

Show universality from model How to show universality from post-updated model?

Show universality from model Use independently control information flow from xx, from yy, from xx by multiplexing forward information from xx and backward information from yy Decomposing WW ii, φφ and error gradient into three parts, WW ii WW ii 0 0 0 WW ii 0, φφ ; θθ ffff, θθ bb 0 0 ww ii φφ ; θθ ffff, θθ bb 00, zz L yy, ff xx; θθ θθ bb 00 ee yy ee yy 2

Show universality from model xx θθ ffff, θθ bb WW NN WW ii WW 1 zz ff oooooo yy xx θθ ffff, θθ bb φφ 00 θθ bb WW NN WW NN ww NN WW ii WW ii ww ii WW 1 WW 1 ww 1 zz zz zz ff oooooo yy

Show universality from model Then, before the gradient update, zz = zz zz zz = NN WW ii φφ(xx; θθ ffff, θθ bb ) ii=1 00 0 Hence, represent the middle one of after gradient update, zz zz = NN ii=1 φφ xx; θθ ffff, θθ bb 00 θθ bb WW ii 0 0 0 WW ii 0 0 0 ww ii NN jj=ii+1 φφ xx; θθ ffff, θθ bb 00 θθ bb WW jj 0 0 0 WW jj 0 0 0 ww jj NN αασ ii=1 ii 1 jj=1 ii 1 jj=1 WW jj 0 0 ii 1 0 WW jj 0 jj=1 0 0 ww jj WW jj 0 0 φφ xx ; θθ ffff, θθ bb 0 WW jj 0 00 0 0 ww jj θθ bb WW jj 0 0 0 WW jj 0 0 0 ww jj 00 ee yy ee yy zz NN = αασ ii=1 ii 1 jj=1 WW jj ii 1 jj=1 WW jj ee yy φφ xx; θθ ffff, θθ bb jj=ii+1 NN ii 1 WW jj jj=1 WW jj φφ xx ; θθ ffff, θθ bb

Show universality from model yy = ff oooooo zz; θθ oooooo and zz = 0 xx θθ ffff, θθ bb φφ 00 θθ bb WW NN WW NN ww NN WW ii WW ii ww ii WW 1 WW 1 ww 1 zz zz zz ff oooooo yy zz NN = αασ ii=1 ii 1 jj=1 WW jj ii 1 jj=1 WW jj ee yy φφ xx; θθ ffff, θθ bb jj=ii+1 = αασ NN ii=1 AA ii ee yy φφ xx; θθ ffff, θθ bb BBii BB ii φφ xx ; θθ ffff, θθ bb (3) NN ii 1 WW jj jj=1 WW jj φφ xx ; θθ ffff, θθ bb xx θθ ffff, θθ bb φφ 00 θθ bb WW NN WW NN ww NN WW ii WW ii ww ii WW 1 WW 1 ww 1 zz zz zz ff oooooo yy

Show universality from model Path 1) Define ff oooooo Path 2) Show zz as new vector vv, which containing informations of xx, yy and xx At last, Universality with Vector vv

Show universality from model Define ff oooooo as a neural network that approximates the following multiplexer function and its derivatives ff oooooo zz zz zz zz ; θθ oooooo = 11 zz=0 gg pppppp ; θθ gg + 11 zz 0 h pppppppp zz; θθ h, 4 zz zz Where gg pppppp is a linear function with parameters θθ gg such that zz L = ee yy (2) Where h pppppppp is a neural network with one or more hidden layers. Then, after pose-update, ff oooooo zz zz zz ; θθ oooooo = h pppppppp zz ; θθ h, 5 xx θθ ffff, θθ bb φφ 00 θθ bb WW NN WW NN ww NN WW ii WW ii ww ii WW 1 WW 1 ww 1 zz zz zz gg pppppp h pppppppp yy

Show universality from model Change the form of zz with kernel form zz = αασ NN ii=1 AA ii ee yy φφ xx; θθ ffff, θθ bb BBii BB ii φφ xx ; θθ ffff, θθ bb = NN αασii=1 AA ii ee yy kk ii xx, xx zz = αασ NN ii=1 AA ii ee yy kk ii xx, xx has the all information of xx, yy and xx. Assume ee yy is a linear function of yy which can extract the original information yy. xx θθ ffff, θθ bb φφ 00 θθ bb WW NN WW NN ww NN WW ii WW ii ww ii WW 1 WW 1 ww 1 zz zz zz gg pppppp h pppppppp yy

Show universality from model Change the form of zz with kernel form zz = αασ NN ii=1 AA ii ee yy φφ xx; θθ ffff, θθ bb BBii BB ii φφ xx ; θθ ffff, θθ bb = NN αασii=1 AA ii ee yy kk ii xx, xx zz = αασ NN ii=1 AA ii ee yy kk ii xx, xx has the all information of xx, yy and xx. Assume ee yy is a linear function of yy which can extract the original information yy. Idea 1) Decompose index ii = 1,, NN to a product of jj = 0,, JJ 1 aaaaaa ll = 0,, LL 1, and then assign those to xx aaaaaa xx Idea 2) Discretize xx and xx to jj and ll by kernel kk(, ) Idea 3) Make new vector vv, which contain all information of xx, xx aaaaaa yy from AA ii, ee yy aaaaaa kk ii

Show universality from model Idea 1) Decompose index ii = 1,, NN to a product of jj = 0,, JJ 1 aaaaaa ll = 0,, LL 1, and then assign those to xx aaaaaa xx zz = αασ NN ii=1 AA ii ee yy kk ii xx, xx = αασ JJ 1 jj=0 Σ ll=0 LL 1 AA jjjj Idea 2) Discretize xx and xx to jj and ll by kernel kk(, ) kk jjjj xx, xx = 1 0 iiii dddddddddd xx = ee jj aaaaaa dddddddddd xx ooooooooooooooooo ee yy kk jjjj xx, xx = ee ll (6) xx θθ ffff, θθ bb φφ 00 θθ bb WW NN WW NN ww NN WW ii WW ii ww ii WW 1 WW 1 ww 1 zz zz zz gg pppppp h pppppppp yy

Show universality from model Idea 3) Make new vector vv, which contain all information of xx, xx aaaaaa yy from AA ii, ee yy aaaaaa kk ii Choose ee yy to be a linear function that outputs J LL stacked copies of yy εε εε Define AA jjjj to be a select yy at the position of (jj, ll) = jj + JJ LL, AA jjjj = 1 + εε, 1 + εε at position (jj, ll) εε εε As a result, zz = αασ NN ii=1 AA ii ee yy kk ii xx, xx 0 0 = vv xx, xx, yy yy 0 0 xx = αασ JJ 1 jj=0 Σ ll=0 θθ ffff, θθ bb LL 1 AA jjjj φφ 00 θθ bb ee yy kk jjjj xx, xx WW NN WW NN ww NN WW ii WW ii ww ii WW 1 WW 1 ww 1 zz zz zz gg pppppp h pppppppp yy

Show universality from model As a result, zz = αασ NN ii=1 AA ii ee yy kk ii xx, xx ff xx ; θθ = αασ JJ 1 jj=0 Σ ll=0 h pppppppp αααα xx, xx, yy ; θθ h LL 1 AA jjjj ee yy kk jjjj xx, xx = αααα xx, xx, yy, where vv xx, xx, yy It is a universal function approximator with respect to its inputs xx, xx, yy 0 0 yy 0 0 xx θθ ffff, θθ bb φφ 00 θθ bb WW NN WW NN ww NN WW ii WW ii ww ii WW 1 WW 1 ww 1 zz zz zz gg pppppp h pppppppp yy

Experiments Q1) Is there empirical benefit to using one meta-learning approach versus another, and in which case? A1) Empirically show the inductive bias of gradient-based vs recurrent meta-learners Explore the differences between gradient-based vs recurrent A learner trained with MAML, improve or start to overfit after additional gradient steps. Better few-shot learning performance on tasks outside of the training distribution?

Experiments

Experiments

Experiments Q2) Theory suggest that deeper networks lead to increased expressive power for representing different learning procedures. Is it right? A1) Investigate the role of model depth in gradient-based metalearning

Experiments

Conclusion Meta-learners that use standard gradient descent with a sufficiently deep representation can approximate any learning procedure, and are equally expressive as recurrent learners. In experiments, MAML is more successful when faced with out-of-domain tasks compared to recurrent model We formalize what it means for a meta-learner to be able to approximate any learning algorithm in terms of its ability to represent functions of the dataset and test inputs.

Appendix A B C we assume inputs and all pre-synaptic activations are non-negative, then deep ReLU act like deep linear networks. (1) zz NN = αασ ii=1 ii 1 jj=1 WW jj ii 1 jj=1 WW jj ee yy φφ xx; θθ ffff, θθ bb jj=ii+1 = αασ NN ii=1 AA ii ee yy φφ xx; θθ ffff, θθ bb BBii BB ii φφ xx ; θθ ffff, θθ bb (3) zz L yy, ff xx; θθ ff oooooo zz zz zz kk jjjj xx, xx = 1 0 00 ee yy ee yy 2 ; θθ oooooo = 11 zz=0 gg pppppp zz zz zz iiii dddddddddd xx = ee jj aaaaaa dddddddddd xx NN ii 1 WW jj jj=1 WW jj h zz; θθ, 4, ff ; θθ gg + 11 zz 0 zz zz pppppppp h oooooo zz ooooooooooooooooo = ee ll (6) φφ xx ; θθ ffff, θθ bb ; θθ oooooo = h pppppppp zz ; θθ h, 5

Appendix A zz NN = αασ ii=1 ii 1 jj=1 WW jj ii 1 jj=1 WW jj ee yy φφ xx; θθ ffff, θθ bb = αασ NN ii=1 AA ii ee yy φφ xx; θθ ffff, θθ bb BBii BB ii φφ xx ; θθ ffff, θθ bb Choose all WW ii and WW ii to be a square and full rank. Set WW ii = MM ii MM 1 ii+1 MM 0 = II) Set AA 1 = II, AA ii = MM ii 1 MM ii 1 1 NN and WW ii = MM ii 1 MM ii then jj=ii+1 and BB ii = MM ii+1, BB NN = II NN jj=ii+1 WW jj ii 1 jj=1 WW jj φφ xx ; θθ ffff, θθ bb ii 1 WW jj = MM ii+1, jj=1 WW jj = MM ii 1 ( MM NN+1 = II,

Appendix A All inputs are non-negative: Input φφ consist of three term, (discretization, constant 0, 0 before update and not used afterward) All inputs are non-negative before and after update. All pre-synaptic activations are non-negative: It is sufficient to show that products of WW ii, ii=jj WW ii (1 jj NN) are positive semi-definite. So, show that NN ii=jj WW ii, NN ii=jj WW ii, NN ii=jj ww ii (1 jj NN) are positive semi-definite. We define NN ii=jj WW ii = MM jj+1 = BB ii, and BB ii is set to be a positive definite. 1 We define WW ii = MM ii 1 MM ii where AA ii = MM ii 1 MM ii 1, so each MM ii is also symmetric positive definite and WW ii is positive definite. Purpose of ww ii is to provide nonzero gradient to the input θθ bb, thus positive value for each ww ii will suffice. NN

Appendix B zz L yy, ff xx; θθ 00 ee yy ee yy 2 zz L yy, ff xx; θθ = zz ff xx; θθ yy L yy, yy = ff xx; θθ = ff oooooo zz; θθ oooooo = gg pppppp zz; θθ gg Because we assume gg pppppp as a linear function of θθ gg, let gg pppppp zz; θθ gg = WW gg WW gg ww gg zz = WW gg zz + WW gg zz gg + ww gg zz To make top element of ee(xx, yy) to be 00, let WW gg = 0, which make yy = 0. Then, ee yy = WW gg yy L yy, 0 For any linear loss function for yy, ee yy = AAAA, to extract all information of yy, AA has to be invertible. Sufficient loss function: standard mean-squared error or softmax cross entropy. 00 ee yy ee yy

Appendix C kk jjjj xx, xx Choose φφ and BB jjjj : = φφ xx; θθ ffff, θθ bb BBjjjj BB jjjj φφ xx ; θθ 1 ffff, θθ bb = 0 φφ ; θθ ffff, θθ bb ddiiiiiiii 0 0 ddiiiiiiii iiff θθ bb = 0 ooooooooooooooooo iiii dddddddddd xx = ee jj aaaaaa dddddddddd xx ooooooooooooooooo BB jjjj = EE jjjj EE jjjj EE llll 0 + εεεε = ee ll Then, where EE iiii denote the matrix a 1 at ii, kk φφ xx; θθ ffff, θθ bb BBjjjj ee jj 0 0 0 iiff dddddddddd xx = ee jj ooooooooooooooooo and 0 otherwise BB jjjj φφ xx ; θθ ee ffff, θθ bb ll 0 iiff dddddddddd xx = ee ll 0 0 ooooooooooooooooo

Appendix C kk jjjj xx, xx = φφ xx; θθ ffff, θθ bb BBjjjj BB jjjj φφ xx ; θθ 1 ffff, θθ bb = 0 φφ ; θθ ffff, θθ bb φφ xx; θθ ffff, θθ bb BBjjjj ee jj 0 0 0 So, ddiiiiiiii 0 0 ddiiiiiiii kk jjjj xx, xx ee jj 0 ee ll 0 0 iiff θθ bb = 0 ooooooooooooooooo iiff dddddddddd xx = ee jj ooooooooooooooooo iiii dddddddddd xx = ee jj aaaaaa dddddddddd xx ooooooooooooooooo BB jjjj = EE jjjj EE jjjj EE llll 0 + εεεε = ee ll BB jjjj φφ xx ; θθ ee ffff, θθ bb ll 0 iiff dddddddddd xx = ee ll 0 0 ooooooooooooooooo iiff dddddddddd xx = ee jj aaaaaa dddddddddd xx ooooooooooooooooo = ee ll

Thank you!