-
Notifications
You must be signed in to change notification settings - Fork 29
/
sb_summit2023.html
444 lines (415 loc) · 30.1 KB
/
sb_summit2023.html
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
<!doctype html>
<html lang="en">
<head>
<!-- Required meta tags -->
<meta charset="utf-8">
<meta name="viewport" content="width=device-width, initial-scale=1, shrink-to-fit=no">
<link rel="icon" href="img/favicon.ico">
<title>About SpeechBrain</title>
<!-- Bootstrap CSS -->
<link rel="stylesheet" href="css/bootstrap.css">
<link rel="stylesheet" href="vendors/linericon/style.css">
<link rel="stylesheet" href="css/font-awesome.min.css">
<link rel="stylesheet" href="vendors/owl-carousel/owl.carousel.min.css">
<link rel="stylesheet" href="vendors/lightbox/simpleLightbox.css">
<link rel="stylesheet" href="vendors/nice-select/css/nice-select.css">
<link rel="stylesheet" href="vendors/animate-css/animate.css">
<!-- main css -->
<link rel="stylesheet" href="css/style.css">
<link rel="stylesheet" href="css/responsive.css">
</head>
<body>
<!--================Header Menu Area =================-->
<header class="header_area">
<div class="main_menu">
<nav class="navbar navbar-expand-lg navbar-light">
<div class="container box_1620">
<!-- Brand and toggle get grouped for better mobile display -->
<a class="navbar-brand logo_h" href="index.html"><img src="img/speechbrain-horiz-logo.svg" width="175px" alt=""></a>
<button class="navbar-toggler" type="button" data-toggle="collapse" data-target="#navbarSupportedContent" aria-controls="navbarSupportedContent" aria-expanded="false" aria-label="Toggle navigation">
<span class="icon-bar"></span>
<span class="icon-bar"></span>
<span class="icon-bar"></span>
</button>
<!-- Collect the nav links, forms, and other content for toggling -->
<div class="collapse navbar-collapse offset" id="navbarSupportedContent">
<ul class="nav navbar-nav menu_nav justify-content-center">
<li class="nav-item active"><a class="nav-link" href="index.html">Home</a></li>
<li class="nav-item"><a class="nav-link" href="about.html">About SpeechBrain</a></li>
<li class="nav-item"><a class="nav-link" href="contributing.html">Contributing</a></li>
<li class="nav-item"><a class="nav-link" href="https://speechbrain.readthedocs.io/">Documentation</a></li>
<li class="nav-item"><a class="nav-link" href="https://speechbrain.readthedocs.io/">Tutorials</a></li>
<li class="nav-item"><a class="nav-link" href="https://github.com/speechbrain/benchmarks">Benchmarks</a></li>
</ul>
</div>
</div>
</nav>
</div>
</header>
<!--================Header Menu Area =================-->
<!--================Home Banner Area =================-->
<section class="home_banner_area blog_banner">
<div class="banner_inner d-flex align-items-center">
<div class="overlay bg-parallax" data-stellar-ratio="0.9" data-stellar-vertical-offset="0" data-background=""></div>
<div class="container">
<div class="blog_b_text text-center">
<img src="img/speechbrain-logo-brain-banner.png" width="90%" style="border-radius: 10%;" alt="">
</div>
</div>
</div>
</section>
<!--================End Home Banner Area =================-->
<!--================Work Area =================-->
<section class="work_area p_120">
<div class="main_title">
<h2>Let’s Talk About Speech Technologies</h2>
<h3>Official recordings are available at <a href="https://www.youtube.com/@SpeechBrainProject/videos">this link</a>!</h3>
<p class="justified large">
SpeechBrain has established itself as a leading deep learning toolkit
for speech processing in recent years, with impressive usage statistics
to back it up. With an average of 100 daily clones and 1000 daily
downloads on its GitHub repository, along with over 6,000 stars and 1100
forks, SpeechBrain is a popular choice among speech processing
experts.
</p>
<br>
<p class="justified large">
In this summit, we are excited to share the latest developments and
updates on the SpeechBrain project, engage in an open and collaborative
discussion with the community, and introduce it to a broader audience of
speech professionals. We would like participants to stay up-to-date with
the latest advancements made in SpeechBrain and speech processing
technology. We also wish to gather, more interactively, the feedback
from the community to better plan future developments. The event will
take place four days after the main conference on August 28th.
</p>
<br>
<p class="justified large"> We are passionate about organizing this event for several reasons: <br/><br/>
<b>•</b> The field of speech technology has seen tremendous growth in recent
years. Our goal is to keep the community informed about the latest
developments and future plans for the SpeechBrain project. We also aim
to engage in an open dialogue with the community to set ambitious goals
for the project’s future. <br/><br/>
<b>•</b> We are excited to bring together experts from both industry and
academia to showcase their impactful projects and share their knowledge
with the community.<br/><br/>
<b>•</b> This event is not only an opportunity to learn and stay updated, but also to network and connect with like-minded individuals in the SpeechBrain community.
The core idea is to avoid missing this chance to be a part of shaping the future of speech technologies and building valuable connections within the community.
</p>
<hr>
<h2>Tentative Schedule</h2>
<p class="justified large"> The event will start at 9am Eastern Daylight Time (EDT) on August 28th and, we will have two sessions with a break from 11.30 am - 12.00 pm. <br/><br/>
<b>Morning (9.00 am - 11.30 am):</b> <br/>
<b>■</b> 9.00 am - 9.30 am <span class="tabulation"></span>Opening and thanks to Sponsors <br/>
<b>■</b> 9.30 am - 10.00 am<span class="tabulation"></span>Industry Talk 1: Arsenii Gorin (UbenwaAI) <br/>
<b>■</b> 10.00 am - 10.30 am<span class="tabulation"></span>Academic Talk 1: Yannick Estève (Université Avignon) <br/>
<b>■</b> 10.30 am - 11.00 am<span class="tabulation"></span>Industry Talk 2: Ariane Nabeth Halber (ViaDialog) <br/>
<b>■</b> 11.00 am - 11.30 am<span class="tabulation"></span>Academic Talk 2: Yan Gao (University of Cambridge) <br/><br/>
<b><i>Lunch Break (11.30 am - 12.00 pm)</b></i> <br/><br/>
<b>Afternoon (12.00 pm - 4.30 pm):</b> <br/>
<b>■</b> 12.00 pm - 1.00 pm<span class="tabulation"></span>SpeechBrain Roadmap 2023 & latest updates <br/>
<b>■</b> 1.00 pm - 1.30 pm<span class="tabulation"></span>Industry Talk 3: Peter Plantinga (JP Morgan Chase) <br/>
<b>■</b> 1.30 pm - 2.00 pm<span class="tabulation"></span>Academic Talk 4: Vielzeuf Valentin (Orange Labs) <br/>
<b>■</b> 2.00 pm - 2.30 pm<span class="tabulation"></span>Coffee Break <br/>
<b>■</b> 2.30 pm - 4.15 pm<span class="tabulation"></span>Panel Discussion and Q&A: <br> Shinji Watanabe, Dan Povey, Brian McFee, Sanchit Gandhi, Zhaoheng Ni <br/>
<b>■</b> 4.15 pm - 4.30 pm<span class="tabulation"></span>Final Remarks and Closing <br/>
</p>
<hr>
<h2>Invited Speakers</h2>
</div>
<div class="speaker-cards">
<div class="speaker-card">
<div class="speaker-image mx-auto">
<img src="img/summit/headshot-peter-plantinga.jpg">
</div>
<div class="speaker-info">
<h3>Peter Plantinga</h3>
<i><h4>JP Morgan Chase & Co</h4></i>
<p class="title">Continual Learning for End-to-End ASR by Averaging Domain Experts</p>
<p>
Peter Plantinga is an Applied AI/ML Associate at the Machine Learning Center of Excellence at JP Morgan Chase & Co. He received his PhD in computer science in 2021 from the Ohio State University (USA) under Prof. Eric Fosler-Lussier focusing on knowledge transfer for the tasks of speech enhancement, robust ASR, and reading verification. His current work involves adapting large-scale ASR models to the financial domain without forgetting, as well as better evaluations of ASR models.
</p>
</div>
</div>
<div class="speaker-card">
<div class="speaker-image mx-auto">
<img src="img/summit/Photo_ArianeNabethHalber_Mars2019_11 - copie.png">
</div>
<div class="speaker-info">
<h3>Ariane Nabeth Halber</h3>
<i><h4>ViaDialog</h4></i>
<p class="title">How ViaDialog sponsors speechbrain and brings hyperlarge vocabulary speech technologies to the contact centres</p>
<p>
Ariane Nabeth-Halber has been working in the speech industry for 25 years. She started her career in research (ATR, Japan; Thalès, France), and then moved to the speech industry, namely with Nuance Communication, and French company Bertin IT, working there with contact centres, broadcasters, trading floors and public ministries, but also academic labs such as LIUM and Avignon University/LIA. Since August 2021, Ariane Nabeth-Halber leads the Speech and Conversational AI team at ViaDialog, to deliver efficient and safe customer relationship experiences. A European Commission expert and LT-Innovate board member, Ariane holds a PhD in computer science and signal processing from Telecom ParisTech. She regularly speaks at conferences on AI and speech technology.
</p>
</div>
</div>
</div>
</div>
<div class="speaker-cards">
<div class="speaker-card">
<div class="speaker-image mx-auto">
<img src="img/summit/arsenii.png">
</div>
<div class="speaker-info">
<h3 class="justified">Arsenii Gorin</h3>
<i><h4>UbenwaAI</h4></i>
<p class="title">Deep learning for infant cry classification</p>
<p class="justified">
Arsenii is the lead ML research scientist Ubenwa. He obtained a PhD from Université de Lorraine working on Automatic Speech Recognition. His main research interests are practical applications of machine learning techniques for audio and speech processing.
</p>
</div>
</div>
<div class="speaker-card">
<div class="speaker-image mx-auto">
<img src="img/summit/vielzeuf.jpg">
</div>
<div class="speaker-info">
<h3 class="justified">Vielzeuf Valentin</h3>
<i><h4>Orange Labs</h4></i>
<p class="title">Speech Recognition Toolkits in Focus: Analyzing Speechbrain's Advantages and Drawbacks through some Orange’s Projects Examples</p>
<p class="justified">
Valentin VIelzeuf is currently a researcher at Orange, focusing on Speech Recognition, Spoken Language Understanding and Complexity Reduction. He holds a phD thesis on Multimodal Deep Learning and also has a background in Computer Vision.
</p>
</div>
</div>
</div>
</div>
<div class="speaker-cards">
<div class="speaker-card">
<div class="speaker-image mx-auto">
<img src="./img/summit/yan_gao.jpg">
</div>
<div class="speaker-info">
<h3 class="justified">Yan Gao</h3>
<i><h4>University of Cambridge</h4></i>
<p class="title">Federated self-supervised speech representation learning</p>
<p class="justified">Yan Gao is a final-year PhD student in Machine Learning System lab at the University of Cambridge, supervised by Prof Nicholas Lane. His research interests are in machine learning, deep learning, optimisation. His recent topics are in federated learning with self-supervised learning on audio and vision data.</p>
</div>
</div>
<div class="speaker-card">
<div class="speaker-image mx-auto">
<img src="./img/summit/yannick22.jpeg">
</div>
<div class="speaker-info">
<h3 class="justified">Yannick Estève</h3>
<i><h4>Avignon Université</h4></i>
<p class="title"> Advancing research: some examples of the SpeechBrain's potential in the context of the LIAvignon partnership chair </p>
<p class="justified">
Yannick received the M.S. (1998) in computer science from the Aix-Marseilles University and the Ph.D. (2002) from Avignon University, France.
He joined Le Mans Université (LIUM lab) in 2003 as an associate professor, and became a full professor in 2010. He moved to Avignon University in 2019 and is the head of the Computer Science Laboratory of Avignon (LIA) since 2020. He has authored and co-authored more than 150 journal and conference papers in speech and language processing. </p>
</div>
</div>
</div>
</div>
<div class="main_title">
<br>
<hr>
<h2>Invited Panel Discussions</h2>
</div>
<div class="speaker-cards">
<div class="speaker-card">
<div class="speaker-image mx-auto">
<img src="img/summit/shinji.jpeg">
</div>
<div class="speaker-info">
<h3 class="justified">Shinji Watanabe</h3>
<i><h4>ESPNet</h4></i>
<p class="justified">
Shinji Watanabe is an Associate Professor at Carnegie Mellon University, Pittsburgh, PA. He received his B.S., M.S., and Ph.D. (Dr. Eng.) degrees from Waseda University, Tokyo, Japan. He was a research scientist at NTT Communication Science Laboratories, Kyoto, Japan, from 2001 to 2011, a visiting scholar at Georgia institute of technology, Atlanta, GA, in 2009, and a senior principal research scientist at Mitsubishi Electric Research Laboratories (MERL), Cambridge, MA USA from 2012 to 2017. Before Carnegie Mellon University, he was an associate research professor at Johns Hopkins University, Baltimore, MD, USA, from 2017 to 2020. His research interests include automatic speech recognition, speech enhancement, spoken language understanding, and machine learning for speech and language processing. He has published over 300 papers in peer-reviewed journals and conferences and received several awards, including the best paper award from the IEEE ASRU in 2019. He is a Senior Area Editor of the IEEE Transactions on Audio Speech and Language Processing. He was/has been a member of several technical committees, including the APSIPA Speech, Language, and Audio Technical Committee (SLA), IEEE Signal Processing Society Speech and Language Technical Committee (SLTC), and Machine Learning for Signal Processing Technical Committee (MLSP). He is an IEEE and ISCA Fellow.</p>
</div>
</div>
<div class="speaker-card">
<div class="speaker-image mx-auto">
<img src="img/summit/danpovey.jpg">
</div>
<div class="speaker-info">
<h3 class="justified">Dan Povey</h3>
<i><h4>Kaldi</h4></i>
<p class="justified">
Daniel Povey is known for many different contributions to the technology of
speech recognition, including early innovations in sequence training such as
Minimum Phone Error, for the Kaldi toolkit, for "next-gen Kaldi" tools
k2/lhotse/Icefall/sherpa, and for the Librispeech dataset.
He completed his PhD at Cambridge University in 2003, spent about
ten years working for industry research labs (IBM Research and then Microsoft
Research) and 7 years as non-tenure-track faculty at Johns Hopkins University.
He moved to Beijing, China in November 2019 to join Xiaomi Corporation as Chief
Voice Scientist.
He is an IEEE Fellow as of 2023.
</p>
</div>
</div>
</div>
</div>
<div class="speaker-cards">
<div class="speaker-card">
<div class="speaker-image mx-auto">
<img src="img/summit/brianmcfee.jpg">
</div>
<div class="speaker-info">
<h3 class="justified">Brian McFee</h3>
<i><h4>Librosa</h4></i>
<p class="justified">
Brian McFee is Assistant Professor of Music Technology and Data Science New York University.
His work lies at the intersection of machine learning and audio analysis.
He is an active open source software developer, and the principal maintainer of the librosa package for audio analysis. </p>
</div>
</div>
<div class="speaker-card">
<div class="speaker-image mx-auto">
<img src="img/summit/sanchit.jpeg">
</div>
<div class="speaker-info">
<h3 class="justified">Sanchit Gandhi</h3>
<i><h4>HuggingFace</h4></i>
<p class="justified">
Sanchit Gandhi is an ML Engineer at Hugging Face. He leads the open-source audio team and maintains the audio models in the Transformers library, with the goal of making state-of-the-art speech recognition models more accessible to the community. Sanchit’s research interests lie in robust, generalisable speech recognition. Prior to working at Hugging Face, Sanchit completed his Masters’ Degree from the University of Cambridge.
</p>
</div>
</div>
</div>
</div>
<div class="speaker-cards">
<div class="speaker-card">
<div class="speaker-image mx-auto">
<img src="img/summit/zhaoheng.png">
</div>
<div class="speaker-info">
<h3 class="justified">Zhaoheng Ni</h3>
<i><h4>Torchaudio</h4></i>
<p class="justified">
Zhaoheng Ni is a research scientist in the PyTorch Audio team of Meta. He graduated from City University of New York supervised by Professor Michael Mandel then joined Meta AI as a research scientist in 2021. His research interests are single-channel and multi-channel speech enhancement, speech separation, and robust ASR. </p>
</div>
</div>
</div>
<div class="main_title">
<br>
<hr>
<h2>Organizing Commitee</h2>
</div>
<div class="speaker-cards">
<div class="speaker-card">
<div class="speaker-image mx-auto">
<img src="img/summit/titouan_parcollet.jpg">
</div>
<div class="speaker-info">
<h3 class="justified">Titouan Parcollet, SpeechBrain, Samsung AI Cambridge.</h3>
<p class="justified">
Titouan is a Research Scientist at the Samsung AI Research center in Cambridge (UK) and a visiting scholar at the Cambridge Machine Learning Systems Lab from the University of Cambridge (UK).
Previously, he was an Associate Professor in computer science at the Laboratoire Informatique d’Avignon (LIA), from Avignon University (FR).
He also was a senior research associate at the University of Oxford (UK) within the Oxford Machine Learning Systems group. He received his PhD in
computer science from the University of Avignon (FR) and in partnership with Orkis focusing on quaternion neural networks, automatic speech recognition,
and representation learning. His current work involves efficient speech recognition, federated learning, and self-supervised learning. He is also currently
collaborating with the University of Montréal (Mila, QC, Canada) as the co-leader of the SpeechBrain project.
</p>
</div>
</div>
<div class="speaker-card">
<div class="speaker-image mx-auto">
<img src="img/summit/cem4_c.jpg">
</div>
<div class="speaker-info">
<h3 class="justified">Cem Subakan, SpeechBrain, Université Laval</h3>
<p class="justified">
Cem is an Assistant Professor at Université Laval in the Computer Science and Software Engineering department.
He is also currently an Affiliate Assistant Professor in the Concordia University Computer Science and Software Engineering Department,
and an invited researcher at Mila, Québec AI Institue. He received his PhD in Computer Science from University of Illinois at Urbana-Champaign (UIUC),
and did a postdoc in Mila Québec AI Institute and Université de Sherbrooke. He serves as reviewer in several conferences including Neurips, ICML, ICLR, ICASSP, MLSP and
journals such as IEEE Signal Processing Letters (SPL), IEEE Transactions on Audio, Speech, and Language Processing (TASL). His research interests include Deep learning for
Source Separation and Speech Enhancement under realistic conditions, Neural Network Interpretability, and Latent Variable Modeling. He is a recipient of best paper award
in the 2017 version IEEE Machine Learning for Signal Processing Conference (MLSP), as well as the Sabura Muroga Fellowship from the UIUC CS department.
He's a core contributor to the SpeechBrain project, leading the speech separation part.
</p>
</div>
</div>
</div>
<div class="speaker-cards">
<div class="speaker-card">
<div class="speaker-image mx-auto">
<img src="img/summit/adel_moumen_2023.png">
</div>
<div class="speaker-info">
<h3 class="justified">Adel Moumen, SpeechBrain, Université d'Avignon</h3>
<p class="justified">
Adel is a research engineer at University of Avignon (FR). He completed his Bachelor's degree in computer science with distinction in an innovation and research-devoted curriculum
and earned a two-year entrepreneurship diploma in 2022. Currently, Adel is participating in a master's apprenticeship program in computer science with specialization in AI,
where he professionally contributes to the development of SpeechBrain, an all-in-one, open-source, PyTorch-based speech processing toolkit. At SpeechBrain, he leads the efforts
of the automatic speech recognition community.
</p>
</div>
</div>
<div class="speaker-card">
<div class="speaker-image mx-auto">
<img src="img/summit/photo_mirco_ravanelli.jpg">
</div>
<div class="speaker-info">
<h3 class="justified">Mirco Ravanelli, SpeechBrain, Concordia University</h3>
<p class="justified">
Mirco is an assistant professor at Concordia University, an adjunct professor at Université de Montréal, and Mila associate member.
His main research interests are deep learning and Conversational AI. He is the author or co-author of more than 60 papers on these research topics.
He received his Ph.D. (with cum laude distinction) from the University of Trento in December 2017.
Mirco is an active member of the speech and machine learning communities. He is the founder and leader of the SpeechBrain project which aim to build an
open-source toolkit for conversational AI and speech processing.
</p>
</div>
</div>
</div>
</div>
<div class="speaker-cards">
<div class="speaker-card">
<div class="speaker-image mx-auto">
<img src="img/summit/francois.jpeg">
</div>
<div class="speaker-info">
<h3 class="justified">François Grondin, SpeechBrain, Université de Sherbrooke</h3>
<p class="justified">
François is an Assistant Professor at Université de Sherbrooke (CA) in the Department of Electrical Engineering and Computer Engineering since 2020.
He was a Postdoctoral Associate at the Computer Science & Artificial Intelligence Laboratory (CSAIL), at Massachusetts Institute of Technology (USA), from 2018 to 2020.
He received his PhD in electrical engineering from Université de Sherbrooke (CA) in 2017. He serves as a reviewer for multiple conferences including ICASSP, INTERSPEECH, RSJ/IROS,
ICRA and journals such as Transactions on Audio, Speech and Language Processing, EURASIP Journal on Audio, Speech, and Music Processing, IEEE Transactions on Robotics, IEEE Robotics and
Automation Letters, IEEE Robotics and Automation Letters, Transactions on Pattern Analysis and Machine Learning. His current work involves multichannel speech enhancement,
sound source localization, ego-noise suppression, sound classification,
robot audition and hybrid signal processing/machine learning approaches. He contributed to the multichannel processing tools in the SpeechBrain project.
</p>
</div>
</div>
<div class="speaker-card">
<div class="speaker-image mx-auto">
<img src="img/summit/andreas.png">
</div>
<div class="speaker-info">
<h3 class="justified">Andreas Nautsch, SpeechBrain, Université d'Avignon</h3>
<p class="justified">
Andreas was research engineer at Université d'Avignon, where he co-maintained SpeechBrain readying it for its next major version release.
Andreas served as project editor for ISO/IEC 19794-13, co-organised editions of the VoicePrivacy and ASVspoof challenges, co-initiated the ISCA SIG on Security & Privacy in Speech Communication (SPSC),
was an Associate Editor for the EURASIP Journal on Audio, Speech, and Music Processing, and co-lead the 2021 Lorentz workshop on Speech as Personal Identifiable Information. By 2020, he lead multidisciplinary
publication teams composed of speech & language technologists, legal scholars, cryptographers, and biometric experts; Andreas co-responded on behalf of ISCA SIG-SPSC to the public consultation of the 2021 EDPB
guidelines on virtual voice assistants. In 2023, he joined goSmart to lead the solution development (architecture, design, and implementation) for private 5G based campus networks.
</p>
</div>
</div>
</div>
</section>
<!--================End Work Area =================-->
<!-- Optional JavaScript -->
<!-- jQuery first, then Popper.js, then Bootstrap JS -->
<script src="js/jquery-3.2.1.min.js"></script>
<script src="js/popper.js"></script>
<script src="js/bootstrap.min.js"></script>
<script src="js/stellar.js"></script>
<script src="vendors/lightbox/simpleLightbox.min.js"></script>
<script src="vendors/nice-select/js/jquery.nice-select.min.js"></script>
<script src="vendors/isotope/imagesloaded.pkgd.min.js"></script>
<script src="vendors/isotope/isotope-min.js"></script>
<script src="vendors/owl-carousel/owl.carousel.min.js"></script>
<script src="js/jquery.ajaxchimp.min.js"></script>
<script src="js/mail-script.js"></script>
<script src="vendors/counter-up/jquery.waypoints.min.js"></script>
<script src="vendors/counter-up/jquery.counterup.min.js"></script>
<script src="js/theme.js"></script>
<link rel="stylesheet" href="js/styles/monokai-sublime.css">
<script src="js/highlight.pack.js"></script>
<script>hljs.initHighlightingOnLoad();</script>
</body>
</html>