| 0:00:06 | okay | 
|---|
| 0:00:07 | so my name is on the money and time from the what a technical to anaemia | 
|---|
| 0:00:12 | and i will | 
|---|
| 0:00:13 | was then you our by uh | 
|---|
| 0:00:15 | its title is analysis of large scale is | 
|---|
| 0:00:17 | i am very not gonna | 
|---|
| 0:00:19 | for language recognition | 
|---|
| 0:00:21 | so | 
|---|
| 0:00:21 | this is the outline of a war | 
|---|
| 0:00:24 | where i was that we don't into that and then i we spend few works on support vector machine | 
|---|
| 0:00:31 | now we discuss some artists | 
|---|
| 0:00:33 | four | 
|---|
| 0:00:33 | five | 
|---|
| 0:00:34 | training overlaps case or vector machine | 
|---|
| 0:00:37 | i will present the subset of that a lot of our and our remote as we | 
|---|
| 0:00:42 | uh trained in order to evaluate the | 
|---|
| 0:00:45 | they are | 
|---|
| 0:00:46 | performances | 
|---|
| 0:00:47 | then i we present our experimental results and then we compute the with some notes um pushed gmms sees them | 
|---|
| 0:00:53 | and the conclusion on on | 
|---|
| 0:00:55 | the | 
|---|
| 0:00:56 | training or something yeah | 
|---|
| 0:00:58 | so | 
|---|
| 0:00:59 | why is | 
|---|
| 0:01:00 | yeah | 
|---|
| 0:01:01 | okay yeah so you can say yeah the svm uh | 
|---|
| 0:01:04 | tend to appear in the many different | 
|---|
| 0:01:06 | every system also necessary but | 
|---|
| 0:01:09 | here we will focus on a lottery system | 
|---|
| 0:01:11 | just to make some it's out on uh | 
|---|
| 0:01:14 | we have | 
|---|
| 0:01:14 | one eight thick anagram based the system G S P S can seize them and pushed gmms | 
|---|
| 0:01:21 | they | 
|---|
| 0:01:21 | are they are quite different but they all share this | 
|---|
| 0:01:25 | that which is the svm training and | 
|---|
| 0:01:27 | classification for phonetic and G S P svm system | 
|---|
| 0:01:31 | in pushing it's actually used in a different light | 
|---|
| 0:01:34 | however they all need some so svm training | 
|---|
| 0:01:38 | so S P N | 
|---|
| 0:01:39 | and support vector machine only not classifier | 
|---|
| 0:01:42 | uh | 
|---|
| 0:01:44 | the objective function on uh can be cast as a regularised the risk minimisation problem | 
|---|
| 0:01:50 | well the loss | 
|---|
| 0:01:51 | for the | 
|---|
| 0:01:54 | most used | 
|---|
| 0:01:54 | the loss function is the hinge loss which use place | 
|---|
| 0:01:58 | to the | 
|---|
| 0:01:59 | what is called a soft margin classifier | 
|---|
| 0:02:02 | and the regularisation term is given by this | 
|---|
| 0:02:05 | where of the normal to the hyperplane | 
|---|
| 0:02:08 | uh which actually is related to the inverse of the margin so we have a trade off between them are | 
|---|
| 0:02:13 | doing | 
|---|
| 0:02:14 | and the misclassification error so | 
|---|
| 0:02:17 | so | 
|---|
| 0:02:19 | uh another formulation is | 
|---|
| 0:02:21 | so given by the dweller grounds and no | 
|---|
| 0:02:23 | the svm problem | 
|---|
| 0:02:26 | uh which is actually a constraint that come with so | 
|---|
| 0:02:29 | musician problem | 
|---|
| 0:02:30 | and this is the culmination is interesting because uh | 
|---|
| 0:02:34 | yeah we have the romantics of dot products between training set bass | 
|---|
| 0:02:40 | and | 
|---|
| 0:02:40 | the fact that we can although just one with the product silos to expand the support vector machine to nonlinear | 
|---|
| 0:02:47 | classification by means of what is called the kernel tree | 
|---|
| 0:02:50 | well we just | 
|---|
| 0:02:52 | my what we choose to wow high data dimensional space by just evaluating dot products in yeah in an evaluation | 
|---|
| 0:03:00 | base without the need | 
|---|
| 0:03:01 | to actually perform any kind of projection | 
|---|
| 0:03:04 | so | 
|---|
| 0:03:05 | well a scalar svm | 
|---|
| 0:03:08 | uh | 
|---|
| 0:03:09 | well uh actually because we have | 
|---|
| 0:03:12 | menu | 
|---|
| 0:03:12 | training part elsewhere yeah larry O nine we have like | 
|---|
| 0:03:15 | seventeen house on which | 
|---|
| 0:03:17 | maybe are not so many | 
|---|
| 0:03:18 | for a recognition system in general but | 
|---|
| 0:03:21 | for our uh | 
|---|
| 0:03:23 | needs | 
|---|
| 0:03:23 | they are | 
|---|
| 0:03:24 | um there are many | 
|---|
| 0:03:26 | and the dimension you are actually baby | 
|---|
| 0:03:29 | 'cause we can go | 
|---|
| 0:03:31 | from | 
|---|
| 0:03:31 | four | 
|---|
| 0:03:32 | the thousand for uh a freedom although with thirty five | 
|---|
| 0:03:35 | Q on it so | 
|---|
| 0:03:37 | to more than one hundred thousand for a gmm system | 
|---|
| 0:03:41 | so uh now we present different targets | 
|---|
| 0:03:44 | to train the | 
|---|
| 0:03:46 | yeah | 
|---|
| 0:03:47 | training | 
|---|
| 0:03:47 | yeah i mean an efficient way | 
|---|
| 0:03:50 | the most of these algorithms are actually uh linear but just for lena cabinets | 
|---|
| 0:03:56 | um | 
|---|
| 0:03:57 | but actually the can we use are almost | 
|---|
| 0:04:01 | always lead us all this | 
|---|
| 0:04:02 | not a problem | 
|---|
| 0:04:04 | so | 
|---|
| 0:04:05 | the this is our baseline system which is lesbian light is one oh | 
|---|
| 0:04:09 | plus files | 
|---|
| 0:04:10 | selena space | 
|---|
| 0:04:11 | solvers | 
|---|
| 0:04:13 | uh it sounds | 
|---|
| 0:04:14 | do a problem uh in any therapy away | 
|---|
| 0:04:17 | and | 
|---|
| 0:04:18 | by decomposing the actual uh problem in smaller subsets | 
|---|
| 0:04:23 | uh the problem with that is that if the | 
|---|
| 0:04:26 | it says that whether i think | 
|---|
| 0:04:28 | time behaviour | 
|---|
| 0:04:29 | so | 
|---|
| 0:04:30 | we could | 
|---|
| 0:04:31 | uh | 
|---|
| 0:04:32 | it has a quality time behaviour and | 
|---|
| 0:04:35 | can this but that uh we did some work to speed it up there | 
|---|
| 0:04:38 | but casing and the gore vacation uh cannot evaluation | 
|---|
| 0:04:43 | uh however | 
|---|
| 0:04:45 | delude a | 
|---|
| 0:04:46 | evaluating on kernel um on the product so | 
|---|
| 0:04:50 | is also whether | 
|---|
| 0:04:52 | problem and the the dramatic | 
|---|
| 0:04:55 | tends to grow | 
|---|
| 0:04:56 | is what i think in the memory | 
|---|
| 0:04:59 | space | 
|---|
| 0:05:00 | so | 
|---|
| 0:05:01 | so we are interested the inadmissible which are actually memory bounded the | 
|---|
| 0:05:06 | and | 
|---|
| 0:05:07 | possibly yeah time on it | 
|---|
| 0:05:10 | so the first one we analyse it was | 
|---|
| 0:05:13 | bag of those which is uh | 
|---|
| 0:05:14 | a primary that is over based on subgradient | 
|---|
| 0:05:18 | all stochastic | 
|---|
| 0:05:19 | grabbed in | 
|---|
| 0:05:20 | the same | 
|---|
| 0:05:21 | well we hear talk about so got in because of the | 
|---|
| 0:05:25 | this function is not actually that immobile everywhere so would | 
|---|
| 0:05:29 | can not | 
|---|
| 0:05:30 | like | 
|---|
| 0:05:30 | the dragon | 
|---|
| 0:05:31 | we | 
|---|
| 0:05:31 | two | 
|---|
| 0:05:32 | subgradient | 
|---|
| 0:05:34 | and we have to cussed selection of learning sample so we do not train every | 
|---|
| 0:05:40 | time | 
|---|
| 0:05:41 | the assistant on the whole database but we just selected randomly | 
|---|
| 0:05:45 | training part | 
|---|
| 0:05:47 | right | 
|---|
| 0:05:48 | you know that | 
|---|
| 0:05:49 | to improve the convergence performance is the | 
|---|
| 0:05:53 | uh the | 
|---|
| 0:05:55 | we have uh projections that on a wall of rogers | 
|---|
| 0:05:58 | so | 
|---|
| 0:05:59 | the square root of the regularisation that we have | 
|---|
| 0:06:02 | the | 
|---|
| 0:06:03 | it's yeah problem formulation | 
|---|
| 0:06:05 | and this is the actually | 
|---|
| 0:06:07 | had | 
|---|
| 0:06:08 | yeah it is | 
|---|
| 0:06:09 | to reach convergence | 
|---|
| 0:06:11 | so this problem the these are gonna do not | 
|---|
| 0:06:14 | that it provides | 
|---|
| 0:06:15 | the | 
|---|
| 0:06:16 | the um | 
|---|
| 0:06:17 | do a solution of the svm problem however | 
|---|
| 0:06:20 | if we need it does | 
|---|
| 0:06:21 | we might do it | 
|---|
| 0:06:23 | we want to implement | 
|---|
| 0:06:24 | jim | 
|---|
| 0:06:24 | pushed gmms as the mighty proposal | 
|---|
| 0:06:27 | uh we can actually train them | 
|---|
| 0:06:30 | why we are trying to be a plane | 
|---|
| 0:06:33 | so | 
|---|
| 0:06:33 | next | 
|---|
| 0:06:34 | we have | 
|---|
| 0:06:35 | do a content inside the this time we move to that was based on | 
|---|
| 0:06:39 | and they can we have an iterative solver which performs | 
|---|
| 0:06:42 | according to this and the interviewer space actually | 
|---|
| 0:06:46 | so we split that one problem uh in uh | 
|---|
| 0:06:49 | uh | 
|---|
| 0:06:50 | in a serious over when evaluate the optimisation so where would keep | 
|---|
| 0:06:55 | all but one variable fixed and we optimise just that one variable by | 
|---|
| 0:07:01 | uh | 
|---|
| 0:07:02 | by some kind of way regarding minimisation | 
|---|
| 0:07:05 | yeah we just have to project regarding you know them | 
|---|
| 0:07:07 | to assure | 
|---|
| 0:07:09 | that | 
|---|
| 0:07:09 | the it's the end what probably constraints that statement | 
|---|
| 0:07:14 | so um | 
|---|
| 0:07:16 | okay uh this time | 
|---|
| 0:07:18 | we do not have | 
|---|
| 0:07:19 | uh that actually the primal solution but actually | 
|---|
| 0:07:23 | it's very easy to update it while we're updating the one solution | 
|---|
| 0:07:28 | and this is nice also because the | 
|---|
| 0:07:30 | uh in order to evaluate the product so we do not have | 
|---|
| 0:07:34 | just or support vector so be careful because we already have the i | 
|---|
| 0:07:38 | plane | 
|---|
| 0:07:39 | so uh this | 
|---|
| 0:07:41 | problem is that it | 
|---|
| 0:07:42 | can be | 
|---|
| 0:07:44 | yeah actually sped up by performing a random permutation of the supplements that is we just switch the order in | 
|---|
| 0:07:51 | which we optimise the variables | 
|---|
| 0:07:54 | and also by introducing some sort of shrieking which are actually means that we | 
|---|
| 0:07:59 | do not | 
|---|
| 0:08:00 | the | 
|---|
| 0:08:01 | uh we tend to not up to update the variables which are | 
|---|
| 0:08:05 | which have region | 
|---|
| 0:08:06 | the bounds of | 
|---|
| 0:08:07 | the guy | 
|---|
| 0:08:07 | the constraints of the svm problem | 
|---|
| 0:08:10 | because uh they will probably | 
|---|
| 0:08:12 | state that we just check | 
|---|
| 0:08:13 | the yeah the this assumption is correct when we meet | 
|---|
| 0:08:17 | so we actually meet | 
|---|
| 0:08:18 | the the comma just | 
|---|
| 0:08:19 | material | 
|---|
| 0:08:22 | so | 
|---|
| 0:08:22 | let's | 
|---|
| 0:08:23 | the | 
|---|
| 0:08:24 | uh we have | 
|---|
| 0:08:24 | it implies some space would so it which was trained | 
|---|
| 0:08:29 | and introduce the in svm pair | 
|---|
| 0:08:32 | this is based on a different formulation of the svm problem the so called ones like | 
|---|
| 0:08:36 | um valuable information | 
|---|
| 0:08:39 | yeah we optimise over the ipod plane and the slack variable | 
|---|
| 0:08:43 | see that | 
|---|
| 0:08:44 | and this time we have a much more greater of a set of | 
|---|
| 0:08:48 | strange so | 
|---|
| 0:08:50 | so what is that what | 
|---|
| 0:08:52 | is that here is that | 
|---|
| 0:08:54 | we have to rebuild that working set about the constraint over which we solve the quadratic problem | 
|---|
| 0:09:01 | and | 
|---|
| 0:09:02 | what's interesting in and | 
|---|
| 0:09:05 | in | 
|---|
| 0:09:05 | these are great | 
|---|
| 0:09:06 | is that the solution is not actually represented by using support back home so | 
|---|
| 0:09:11 | but as a present the but means so what they call basis vector which are essentially have the same role | 
|---|
| 0:09:17 | but they are not actually taken from the training set itself | 
|---|
| 0:09:21 | so uh what we obtain is that we have uh much | 
|---|
| 0:09:25 | sparser representation because the number of basis vectors this much | 
|---|
| 0:09:31 | i'm only meet at the with respect to the support vector which actually tend to increase in number | 
|---|
| 0:09:37 | and uh linearly with the size with the training set size | 
|---|
| 0:09:41 | however the problem is that this time | 
|---|
| 0:09:43 | cannot | 
|---|
| 0:09:44 | assuming | 
|---|
| 0:09:45 | blair recovered one solution of the svm problem | 
|---|
| 0:09:50 | uh what is nice to actually see that | 
|---|
| 0:09:52 | since we have uh so few buttons back over | 
|---|
| 0:09:55 | it is | 
|---|
| 0:09:56 | is it to extend this technique to an only not cameras | 
|---|
| 0:10:00 | but actually we didn't try | 
|---|
| 0:10:03 | so find the final and i agree | 
|---|
| 0:10:04 | yeah is the | 
|---|
| 0:10:06 | yeah that ma'am this is that they can from my uh risk minimisation framework | 
|---|
| 0:10:11 | uh which are uh as that of the for the svm is a what | 
|---|
| 0:10:15 | clotting | 
|---|
| 0:10:16 | and so | 
|---|
| 0:10:17 | so this time again we build any domain that low we won't be at work | 
|---|
| 0:10:22 | set of approximate solution by taking tangent planes | 
|---|
| 0:10:26 | objective function | 
|---|
| 0:10:28 | and actually solving and ah | 
|---|
| 0:10:31 | the minimisation on | 
|---|
| 0:10:32 | these up | 
|---|
| 0:10:33 | on the the functional approximated by means of pungent planes | 
|---|
| 0:10:38 | so this time we still need to solve a a quadratic problem but | 
|---|
| 0:10:42 | the size of the quality problem is actually | 
|---|
| 0:10:45 | motion and i actually equal to the number of uh | 
|---|
| 0:10:49 | tangent plane so we are using to approximate the function | 
|---|
| 0:10:52 | and this is also equal to the in number of iteration with taking so | 
|---|
| 0:10:58 | the size of this problem is much much smaller than the size of the original problem | 
|---|
| 0:11:03 | and usually can be neglected since we do not | 
|---|
| 0:11:06 | the need | 
|---|
| 0:11:07 | more than two hundred or something like that | 
|---|
| 0:11:10 | iterations | 
|---|
| 0:11:11 | so i can be | 
|---|
| 0:11:12 | hundreds of the | 
|---|
| 0:11:14 | hmmm | 
|---|
| 0:11:14 | the primal formulation of the svm problem but | 
|---|
| 0:11:18 | do a solution can be about right | 
|---|
| 0:11:19 | the is the also | 
|---|
| 0:11:21 | this time | 
|---|
| 0:11:23 | oh | 
|---|
| 0:11:23 | now yeah they re more than we try to | 
|---|
| 0:11:26 | uh larry model without a doubt a small subset of | 
|---|
| 0:11:30 | the more the so we use what you larry oh no no evaluation | 
|---|
| 0:11:34 | it's not the phonetic model is just | 
|---|
| 0:11:36 | us under the | 
|---|
| 0:11:38 | and bigram based the system when we perform connected according using a italian | 
|---|
| 0:11:43 | tokenizer | 
|---|
| 0:11:44 | then we're stuck and they don't count so we perform svm training and we adopt | 
|---|
| 0:11:49 | that yeah but a lot of uh | 
|---|
| 0:11:51 | kennel uh uh which actually is really not cannot so we just | 
|---|
| 0:11:55 | to perform some kind of | 
|---|
| 0:11:57 | twenty but the normalisation before feeding them to the svm on it | 
|---|
| 0:12:02 | then the acoustic system is a standard two thousand forty eight gosh amount | 
|---|
| 0:12:06 | well that the the gym and uh you | 
|---|
| 0:12:09 | six parameters | 
|---|
| 0:12:10 | we | 
|---|
| 0:12:12 | we study the gaussian means two things | 
|---|
| 0:12:14 | supervectors and we use the | 
|---|
| 0:12:16 | okay i can do that | 
|---|
| 0:12:18 | again just | 
|---|
| 0:12:19 | normalising | 
|---|
| 0:12:21 | the | 
|---|
| 0:12:21 | buttons | 
|---|
| 0:12:22 | so the system we actually train it is uh | 
|---|
| 0:12:26 | we were interested in evaluating | 
|---|
| 0:12:27 | was | 
|---|
| 0:12:28 | gmm push system | 
|---|
| 0:12:30 | where we used the svm meant what solution as the combination weights | 
|---|
| 0:12:35 | for the model and the T model | 
|---|
| 0:12:37 | and | 
|---|
| 0:12:38 | so | 
|---|
| 0:12:39 | no uh scoring is performed by means of a lack of duration | 
|---|
| 0:12:44 | oh the evaluation condition we tested it on the L every O nine which combines | 
|---|
| 0:12:49 | twenty three languages with narrowband broadcast and telephone data | 
|---|
| 0:12:54 | we tested the systems on the thirty second and second and three second the valuation conditions | 
|---|
| 0:12:59 | training was performed using seven | 
|---|
| 0:13:01 | T in a thousand more or less training sentences | 
|---|
| 0:13:05 | uh the main difference between what we did this time and what we did in the end larry O nine | 
|---|
| 0:13:10 | evaluation is that | 
|---|
| 0:13:12 | this then we train channel independent | 
|---|
| 0:13:14 | system | 
|---|
| 0:13:15 | well | 
|---|
| 0:13:16 | for yeah that realign we use | 
|---|
| 0:13:17 | channel dependent system | 
|---|
| 0:13:19 | so on more that's not training you know one buttons or party | 
|---|
| 0:13:24 | class balancing is | 
|---|
| 0:13:25 | simulated the for all systems | 
|---|
| 0:13:27 | except for a svm pad which do not actually allow for easy simulation of class balancing we | 
|---|
| 0:13:34 | we performed by just playing with this | 
|---|
| 0:13:37 | see fat or uh | 
|---|
| 0:13:39 | and the | 
|---|
| 0:13:40 | and the loss function | 
|---|
| 0:13:42 | and in order to improve uh uh time performance is all models when training together | 
|---|
| 0:13:49 | that you know to to just | 
|---|
| 0:13:51 | have to scan one uh | 
|---|
| 0:13:53 | with just one second of the database we train | 
|---|
| 0:13:56 | uh we we trained on the more than some | 
|---|
| 0:13:59 | so | 
|---|
| 0:14:00 | here are the results | 
|---|
| 0:14:02 | the four | 
|---|
| 0:14:04 | the phonetic system okay | 
|---|
| 0:14:06 | is the is | 
|---|
| 0:14:07 | just | 
|---|
| 0:14:07 | the | 
|---|
| 0:14:08 | the same | 
|---|
| 0:14:09 | system using uh this way or | 
|---|
| 0:14:12 | uh the hinge loss function | 
|---|
| 0:14:14 | uh with this idea but | 
|---|
| 0:14:16 | you know | 
|---|
| 0:14:17 | true | 
|---|
| 0:14:17 | two | 
|---|
| 0:14:18 | to give it any kind of use | 
|---|
| 0:14:21 | one uh results | 
|---|
| 0:14:22 | okay so what do you can see here is that actually | 
|---|
| 0:14:26 | uh the | 
|---|
| 0:14:27 | all results when all the system and which have made the | 
|---|
| 0:14:32 | type | 
|---|
| 0:14:32 | the combatants | 
|---|
| 0:14:34 | criterion | 
|---|
| 0:14:35 | so uh we can see that | 
|---|
| 0:14:37 | although the the assistant performs | 
|---|
| 0:14:39 | almost the same except for svm pair which is uh | 
|---|
| 0:14:43 | due to the lack of class about nothing | 
|---|
| 0:14:46 | uh so this is the same for the acoustic system | 
|---|
| 0:14:50 | uh the results are almost the same | 
|---|
| 0:14:52 | this time he were here we do not they're svm back | 
|---|
| 0:14:55 | 'cause | 
|---|
| 0:14:56 | we do not have the the | 
|---|
| 0:14:58 | and uh that was solution in uh we need | 
|---|
| 0:15:01 | right | 
|---|
| 0:15:01 | the push gmm system | 
|---|
| 0:15:04 | so uh i mean was that so okay here is the the phonetic | 
|---|
| 0:15:08 | so the second condition | 
|---|
| 0:15:10 | uh what | 
|---|
| 0:15:11 | we can see is that actually the C D M performs very well | 
|---|
| 0:15:16 | yeah uh svm like which is our baseline is not shown because it to like | 
|---|
| 0:15:21 | more than nine thousand seconds to train it so we just | 
|---|
| 0:15:25 | but does show in the D C S | 
|---|
| 0:15:27 | but | 
|---|
| 0:15:28 | result but | 
|---|
| 0:15:29 | the time | 
|---|
| 0:15:31 | it to just didn't | 
|---|
| 0:15:33 | there was just too much to show it in the plot | 
|---|
| 0:15:36 | and so we | 
|---|
| 0:15:37 | we can see is that | 
|---|
| 0:15:38 | actually all the arguments the | 
|---|
| 0:15:41 | yeah | 
|---|
| 0:15:43 | improved performance with three time but | 
|---|
| 0:15:46 | the more | 
|---|
| 0:15:47 | for me one is actually record in the senate which allows us | 
|---|
| 0:15:51 | train the svm in less | 
|---|
| 0:15:53 | then | 
|---|
| 0:15:54 | more or less to handle seconds | 
|---|
| 0:15:56 | so this is the same for the test is then condition | 
|---|
| 0:16:00 | and the three system condition actually here | 
|---|
| 0:16:03 | we can not so | 
|---|
| 0:16:04 | that | 
|---|
| 0:16:04 | so | 
|---|
| 0:16:05 | uh the generalisation um | 
|---|
| 0:16:08 | generalisation uh property of the svm and | 
|---|
| 0:16:12 | is actually | 
|---|
| 0:16:13 | as is like | 
|---|
| 0:16:14 | back | 
|---|
| 0:16:15 | note before we | 
|---|
| 0:16:16 | actually reach the the convergence criterion | 
|---|
| 0:16:19 | however it's | 
|---|
| 0:16:20 | not so relevant | 
|---|
| 0:16:22 | so | 
|---|
| 0:16:23 | now we get to the acoustic | 
|---|
| 0:16:25 | pushed gmms is yeah | 
|---|
| 0:16:26 | okay but the effect on the north in | 
|---|
| 0:16:30 | nothing you with respect to the previews the grass | 
|---|
| 0:16:34 | uh we also have the the city average | 
|---|
| 0:16:37 | forms | 
|---|
| 0:16:37 | quite well | 
|---|
| 0:16:38 | well there is just a small difference in terms of this year | 
|---|
| 0:16:42 | between as to seize them and then one sees them but | 
|---|
| 0:16:45 | it's | 
|---|
| 0:16:46 | just | 
|---|
| 0:16:47 | very little | 
|---|
| 0:16:48 | so uh for the ten second condition with | 
|---|
| 0:16:52 | start to see some interesting things | 
|---|
| 0:16:55 | so | 
|---|
| 0:16:55 | and | 
|---|
| 0:16:56 | actually what we get | 
|---|
| 0:16:58 | here is that | 
|---|
| 0:16:59 | the statistical condition we obtain results | 
|---|
| 0:17:02 | which we | 
|---|
| 0:17:02 | did not expect | 
|---|
| 0:17:04 | so here for pushes them as to what | 
|---|
| 0:17:07 | you can see is that actually | 
|---|
| 0:17:09 | the | 
|---|
| 0:17:10 | okay you | 
|---|
| 0:17:12 | the first part of the graph here | 
|---|
| 0:17:14 | represents using pushing way so which are far from the svm optimum | 
|---|
| 0:17:18 | so what we obtain is that actually using the svm optimum uh | 
|---|
| 0:17:24 | uh is not optimal for uh at least for the three second condition for pushed gmms | 
|---|
| 0:17:29 | uh so we can see that | 
|---|
| 0:17:31 | to | 
|---|
| 0:17:32 | yeah we have the first iteration of the bbn matter | 
|---|
| 0:17:35 | be an atom heart body | 
|---|
| 0:17:37 | which i study is like we wanted | 
|---|
| 0:17:39 | two | 
|---|
| 0:17:40 | but for pushing by simply taking the arithmetical mean of | 
|---|
| 0:17:44 | true class time post and the fourth circuit same posted up | 
|---|
| 0:17:47 | any need of a svm training | 
|---|
| 0:17:49 | and actually for the three second condition is | 
|---|
| 0:17:52 | in this performs even better they're using | 
|---|
| 0:17:55 | the outer door waiting | 
|---|
| 0:17:57 | and | 
|---|
| 0:17:58 | okay here we have uh | 
|---|
| 0:18:01 | some other kinds of weighting which is | 
|---|
| 0:18:03 | very far from the optimal actually we're | 
|---|
| 0:18:06 | they have no or not | 
|---|
| 0:18:08 | no i understand the board meeting this way | 
|---|
| 0:18:10 | the just | 
|---|
| 0:18:12 | around them intonational the algorithm however they | 
|---|
| 0:18:15 | they are very far from the optimal | 
|---|
| 0:18:17 | but they are | 
|---|
| 0:18:18 | the best performing system | 
|---|
| 0:18:20 | for the second condition | 
|---|
| 0:18:22 | so | 
|---|
| 0:18:22 | yeah is what those | 
|---|
| 0:18:24 | what they said don't form pushed gmms is then we obtain the with the results | 
|---|
| 0:18:29 | even when the the push away some very far from the svm optimum | 
|---|
| 0:18:34 | uh model and then to more than the train just taking the arithmetic mean on the thirty second condition actually | 
|---|
| 0:18:41 | improves performance is a | 
|---|
| 0:18:43 | and well let's | 
|---|
| 0:18:44 | so while svm push a base pushing | 
|---|
| 0:18:48 | improves performance for the thirty seconds and slightly for this ten seconds condition | 
|---|
| 0:18:54 | for the three seconds condition we actually look | 
|---|
| 0:18:56 | have to look | 
|---|
| 0:18:58 | so now the conclusion on svm modelling | 
|---|
| 0:19:01 | we trained different are great | 
|---|
| 0:19:03 | no | 
|---|
| 0:19:04 | and uh what we obtain is that actually this at the emmys | 
|---|
| 0:19:07 | process one themselves | 
|---|
| 0:19:09 | well problems so we if we are still interested in the the solution these | 
|---|
| 0:19:13 | is provided but is that really | 
|---|
| 0:19:16 | however uh design put it in other wise | 
|---|
| 0:19:19 | at this the | 
|---|
| 0:19:21 | the the svm solution after | 
|---|
| 0:19:24 | they're each button so | 
|---|
| 0:19:26 | not | 
|---|
| 0:19:26 | right | 
|---|
| 0:19:28 | cannot directly be that's why the in uh this a good or a class to another minor mental | 
|---|
| 0:19:34 | so i svm powerful is the second fastest are great no and | 
|---|
| 0:19:38 | can | 
|---|
| 0:19:39 | take advantage over this to put on the environment since um they separate from just at the end of one | 
|---|
| 0:19:44 | and uh of a complete database can | 
|---|
| 0:19:47 | and the scaling is good also for normally not can or however the | 
|---|
| 0:19:52 | do what solution is not provided | 
|---|
| 0:19:54 | and | 
|---|
| 0:19:55 | class balancing can not be directly implemented the in an easy way out | 
|---|
| 0:19:59 | it is possible with the other art | 
|---|
| 0:20:02 | then B R M and uh B M a time is a much slower than the other arguments | 
|---|
| 0:20:07 | however | 
|---|
| 0:20:08 | we still have to see how how far we can how much we can speed up by using distributed environment | 
|---|
| 0:20:15 | since | 
|---|
| 0:20:15 | again this time | 
|---|
| 0:20:17 | uh | 
|---|
| 0:20:18 | weights um solution update is performed by a computer after that a complete database can | 
|---|
| 0:20:24 | and also what is interesting and is a great it that's it | 
|---|
| 0:20:28 | it's better is it works | 
|---|
| 0:20:29 | then | 
|---|
| 0:20:30 | uh what we did | 
|---|
| 0:20:31 | to do | 
|---|
| 0:20:32 | very different loss functions | 
|---|
| 0:20:34 | and finally the inspectors | 
|---|
| 0:20:36 | which which through the slower than the other one is the more already | 
|---|
| 0:20:40 | can not | 
|---|
| 0:20:42 | exploited it's a bit on the environment | 
|---|
| 0:20:44 | so | 
|---|
| 0:20:45 | like the | 
|---|
| 0:20:46 | uh that was | 
|---|
| 0:20:47 | so | 
|---|
| 0:20:48 | before | 
|---|
| 0:20:49 | so | 
|---|
| 0:20:49 | this is all | 
|---|
| 0:20:51 | uh | 
|---|
| 0:20:51 | questions | 
|---|
| 0:20:59 | exactly | 
|---|
| 0:20:59 | your question | 
|---|
| 0:21:06 | since then | 
|---|
| 0:21:07 | actually | 
|---|
| 0:21:09 | some of your | 
|---|
| 0:21:10 | uh | 
|---|
| 0:21:11 | it's yeah | 
|---|
| 0:21:12 | summarising | 
|---|
| 0:21:14 | oh great | 
|---|
| 0:21:15 | along the way | 
|---|
| 0:21:15 | fine | 
|---|
| 0:21:16 | oh | 
|---|
| 0:21:17 | oh | 
|---|
| 0:21:18 | yeah | 
|---|
| 0:21:18 | forming | 
|---|
| 0:21:19 | yeah solution | 
|---|
| 0:21:20 | does it mean that actually | 
|---|
| 0:21:21 | yeah | 
|---|
| 0:21:22 | roger | 
|---|
| 0:21:23 | the model | 
|---|
| 0:21:26 | right | 
|---|
| 0:21:27 | like | 
|---|
| 0:21:29 | right | 
|---|
| 0:21:30 | yeah | 
|---|
| 0:21:31 | the different you | 
|---|
| 0:21:33 | well | 
|---|
| 0:21:34 | okay | 
|---|
| 0:21:35 | actually i think that | 
|---|
| 0:21:37 | a uh | 
|---|
| 0:21:38 | the svm problem uh the optimal solution actually is trying to minimise | 
|---|
| 0:21:43 | the estimation of yeah | 
|---|
| 0:21:45 | of the general | 
|---|
| 0:21:46 | is that just that estimating the generalisation error or the svm uh | 
|---|
| 0:21:50 | because we training it on train set and it tries to estimate the generalisation error but | 
|---|
| 0:21:55 | still using the training set | 
|---|
| 0:21:57 | so when we actually deployed | 
|---|
| 0:21:59 | so it | 
|---|
| 0:22:00 | it that can solve than that | 
|---|
| 0:22:02 | the actual uh best generalisation error is not the thing that when we have | 
|---|
| 0:22:08 | reached | 
|---|
| 0:22:09 | type combat and the materials but | 
|---|
| 0:22:11 | when we poses | 
|---|
| 0:22:14 | uh less | 
|---|
| 0:22:15 | five | 
|---|
| 0:22:15 | uh criterion for convergence | 
|---|
| 0:22:18 | just at some iteration | 
|---|
| 0:22:20 | before | 
|---|
| 0:22:20 | actually produce it | 
|---|
| 0:22:21 | this again | 
|---|
| 0:22:22 | in some iteration before | 
|---|
| 0:22:24 | whatever | 
|---|
| 0:22:25 | uh maybe it's just | 
|---|
| 0:22:28 | imposing like | 
|---|
| 0:22:30 | less tighter conditions | 
|---|
| 0:22:38 | not to mention | 
|---|
| 0:22:40 | uh | 
|---|
| 0:22:42 | what we did | 
|---|
| 0:22:45 | training process | 
|---|
| 0:22:47 | let's | 
|---|
| 0:22:47 | then yeah | 
|---|
| 0:22:49 | maybe | 
|---|
| 0:22:50 | nominee | 
|---|
| 0:22:50 | anyway | 
|---|
| 0:22:51 | you know the | 
|---|
| 0:22:54 | your | 
|---|
| 0:22:56 | the number of samples that you trained on most uh | 
|---|
| 0:22:58 | entails | 
|---|
| 0:22:59 | yeah | 
|---|
| 0:23:00 | your | 
|---|
| 0:23:00 | oh | 
|---|
| 0:23:01 | kernel matrix | 
|---|
| 0:23:02 | not not | 
|---|
| 0:23:03 | she said | 
|---|
| 0:23:04 | min | 
|---|
| 0:23:05 | yeah yeah well actually it should | 
|---|
| 0:23:07 | but | 
|---|
| 0:23:08 | and uh we went | 
|---|
| 0:23:09 | from a lower rio five where we have five thousand to everyone and where we have seventeen thousand so we | 
|---|
| 0:23:15 | don't know next to what we have uh | 
|---|
| 0:23:18 | so if we would have liked | 
|---|
| 0:23:19 | thirty thousand and uh i don't know if | 
|---|
| 0:23:22 | it | 
|---|
| 0:23:23 | just | 
|---|
| 0:23:23 | i | 
|---|
| 0:23:24 | memory | 
|---|
| 0:23:25 | so | 
|---|
| 0:23:26 | yeah yeah actually uh S P and i was trained by evaluating the uh can on my tricks and storing | 
|---|
| 0:23:32 | it in my memory | 
|---|
| 0:23:35 | i where you | 
|---|
| 0:23:37 | perform | 
|---|
| 0:23:37 | or send you it | 
|---|
| 0:23:44 | i had a quick question when you say class balancing you mean uh | 
|---|
| 0:23:48 | giving equal to the weight | 
|---|
| 0:23:50 | so the | 
|---|
| 0:23:50 | positive and the negative yeah well actually i mean we tried to simulate where the same number of true samples | 
|---|
| 0:23:57 | four samples by playing with the uh | 
|---|
| 0:24:00 | but i mean | 
|---|
| 0:24:00 | with the scaling parameter of the loss function | 
|---|
| 0:24:03 | but is | 
|---|
| 0:24:04 | we just | 
|---|
| 0:24:05 | divide the through loss | 
|---|
| 0:24:08 | the to the losses for the two buttons and the losses for the most part | 
|---|
| 0:24:12 | just wait and if | 
|---|
| 0:24:13 | yeah | 
|---|
| 0:24:15 | okay let's take the speaker can | 
|---|