-
Notifications
You must be signed in to change notification settings - Fork 0
/
feed.xml
242 lines (185 loc) · 14.9 KB
/
feed.xml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
<?xml version="1.0" encoding="UTF-8" ?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
<channel>
<title>A Not-So Primordial Soup</title>
<link>https://blog.quipu-strands.com</link>
<atom:link href="https://blog.quipu-strands.com/feed.xml" rel="self" type="application/rss+xml" />
<description>My thought-recorder.</description>
<lastBuildDate>Sun, 29 Sep 2024 12:32:36 -0700</lastBuildDate>
<item>
<title>Inactive Learning?</title>
<link>
https://blog.quipu-strands.com/inactive_learning
</link>
<description>
<script type="text/x-mathjax-config">
MathJax.Hub.Config({
"HTML-CSS": { scale: 100, linebreaks: { automatic: true } },
SVG: { linebreaks: { automatic:true } },
displayAlign: "center" });
</script>
<script type="text/javascript" async="" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML">
</script>
<p>I totally stole the title from a paper <a class="citation" href="#10.1145/1964897.1964906">(Attenberg &amp; Provost, 2011)</a>.</p>
<p>In theory, <em>Active Learning (AL)</em> is a tremendous idea. You need labeled data, but your kind of labeling comes at a cost, e.g., you need to obtain them from a domain expert. Now, lets say, your goal is to use this labeled data to train a classifier that gets to a held-out accuracy of \(90\%\). If you randomly sampled points to label, you might require \(1000\) points. Active Learning lets you <em>strategically</em> pick just \(500\) points for labeling, to reach the same accuracy. Half the labeling cost for the same outcome. This is great!</p>
<p>Except that in a lot of real-world cases this is not how it plays out. I suspected this from my personal experiments, and then in some stuff we did at <a href="https://www.247.ai/">[24]7.ai</a>. So we decided to thoroughly test out multiple scenarios in text classification, where you believe (or current literature leads us to believe) Active Learning <em>should</em> work … but it just doesn’t. We summarized our observations into the paper <em>“On the Fragility of Active Learners for Text Classification”</em> <a class="citation" href="#fragilityActive">(Ghose &amp; Nguyen, 2024)</a> [<a href="https://arxiv.org/pdf/2403.15744">PDF</a>], and th...
</description>
<pubDate>Wed, 25 Sep 2024 12:00:00 -0700</pubDate>
<guid>
https://blog.quipu-strands.com/inactive_learning
</guid>
</item>
<item>
<title>Jensen's Inequality - A Visual Intuition</title>
<link>
https://blog.quipu-strands.com/jensens_inequality
</link>
<description>
<script type="text/x-mathjax-config">
MathJax.Hub.Config({
"HTML-CSS": { scale: 100, linebreaks: { automatic: true } },
SVG: { linebreaks: { automatic:true } },
displayAlign: "center" });
</script>
<script type="text/javascript" async="" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML">
</script>
<p><strong>Jensen’s inequality</strong> finds widespread application in mathematical proofs. I am fond of a particular intuitive explanation of it, which doesn’t seem to be very popular. I will try to present it in brief here.</p>
<p>I am not sure when this argument originated, but Google does turn up a paper <a class="citation" href="#jensens_needham">(Needham, 1993)</a>. Even if this is not the source, it is a good reference. On a related note, the author of the paper, Tristan Needham, has the very well-reviewed book “<a href="https://www.amazon.com/Visual-Complex-Analysis-Tristan-Needham/dp/0198534469">Visual Complex Analysis</a>” to his credit.</p>
<p>For the record, <a href="(https://en.wikipedia.org/wiki/Jensen%27s_inequality)">Wikipedia</a> states one version in this way:</p>
<!-- _includes/image.html -->
<div class="image-wrapper">
<img src="/assets/jensens/wiki.png" alt="test" />
<p class="image-caption">Jensen's Inequality, as on Wikipedia.</p>
</div>
<p>But don’t read too much into this yet; we’ll discover this form along the way.
This is going to sound surprising but let’s start with the idea of the <em>center of mass (CM)</em>.</p>
<h2 id="center-of-mass">Center of Mass</h2>
<p>For our purposes, we can think of th...
</description>
<pubDate>Sun, 15 Sep 2024 23:44:00 -0700</pubDate>
<guid>
https://blog.quipu-strands.com/jensens_inequality
</guid>
</item>
<item>
<title>Bayesian Optimization, Part 2: Acquisition Functions</title>
<link>
https://blog.quipu-strands.com/bayesopt_2_acq_fns
</link>
<description>
<script type="text/x-mathjax-config">
MathJax.Hub.Config({
"HTML-CSS": { scale: 100, linebreaks: { automatic: true } },
SVG: { linebreaks: { automatic:true } },
displayAlign: "center" });
</script>
<script type="text/javascript" async="" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML">
</script>
<!-- _includes/image.html -->
<div class="image-wrapper">
<img src="/assets/bayesopt/bayesopt_cover.png" alt="test" />
<p class="image-caption"></p>
</div>
<p>This post continues our discussion on BayesOpt. This is <strong>part-2 of a two-part series</strong>.
Now we take a look at the other pillar BayesOpt rests on: acquisition functions. My goal is to provide a flavor by looking at a few of them. I’ll go into depth for a couple; this would help us appreciate the role of GPs in conveniently calculating acquisition values. For the rest I’ll provide an overview.</p>
<ul id="markdown-toc">
<li><a href="#acquisition-functions" id="markdown-toc-acquisition-functions">Acquisition Functions</a> <ul>
<li><a href="#probability-of-improvement" id="markdown-toc-probability-of-improvement">Probability of Improvement</a></li>
<li><a href="#expected-improvement" id="markdown-toc-expected-improvement">Expected Improvement</a></li>
<li><a href="#predictive-entropy-search" id="markdown-toc-predictive-entropy-search">Predictive Entropy Search</a></li>
<li><a href="#max-value-entropy-search" id="markdown-toc-max-value-entropy-search">Max-value Entropy ...
</description>
<pubDate>Sat, 18 Nov 2023 11:00:00 -0800</pubDate>
<guid>
https://blog.quipu-strands.com/bayesopt_2_acq_fns
</guid>
</item>
<item>
<title>Bayesian Optimization, Part 1: Key Ideas, Gaussian Processes</title>
<link>
https://blog.quipu-strands.com/bayesopt_1_key_ideas_GPs
</link>
<description>
<script type="text/x-mathjax-config">
MathJax.Hub.Config({
"HTML-CSS": { scale: 100, linebreaks: { automatic: true } },
SVG: { linebreaks: { automatic:true } },
displayAlign: "center" });
</script>
<script type="text/javascript" async="" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML">
</script>
<!-- _includes/image.html -->
<div class="image-wrapper">
<img src="/assets/bayesopt/bayesopt_cover.png" alt="test" />
<p class="image-caption">The real reason I like Bayesian Optimization: lots of pretty pictures!</p>
</div>
<p>If I wanted to sell you on the idea of <em>Bayesian Optimization (BayesOpt)</em>, I’d just list some of its applications:</p>
<ul>
<li>Hyperparameter Optimization (HPO) <a class="citation" href="#bayesopt_is_superior">(Turner et al., 2021)</a>.</li>
<li>Neural Architecture Search (NAS) <a class="citation" href="#white2019bananas">(White et al., 2021)</a>.</li>
<li>Molecule discovery <a class="citation" href="#LSBO_paper">(Gómez-Bombarelli et al., 2018)</a>.</li>
<li>Liquid chromatography <a class="citation" href="#BOELRIJK2023340789">(Boelrijk et al., 2023)</a>.</li>
<li>Creating low-carbon concrete <a class="citation" href="#ament2023sustainable">(Ament et al., 2023)</a>.</li>
<li>Plasma control in nuclear fusion <a class="citation" href="#mehta2022an">(Mehta et al., 2022)</a>.</li>
<li>Parameter tuning for lasers <a class="citation" href="#20.500.11850/385...
</description>
<pubDate>Sat, 18 Nov 2023 11:00:00 -0800</pubDate>
<guid>
https://blog.quipu-strands.com/bayesopt_1_key_ideas_GPs
</guid>
</item>
<item>
<title>Fun with GMMs</title>
<link>
https://blog.quipu-strands.com/fun_with_GMMs
</link>
<description>
<script type="text/x-mathjax-config">
MathJax.Hub.Config({
"HTML-CSS": { scale: 100, linebreaks: { automatic: true } },
SVG: { linebreaks: { automatic:true } },
displayAlign: "center" });
</script>
<script type="text/javascript" async="" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML">
</script>
<p><em>Generative Models</em> have been all the rage in AI lately, be it image generators like <a href="https://stability.ai/blog/stable-diffusion-public-release">Stable Diffusion</a> or text generators like <a href="https://openai.com/blog/chatgpt/">ChatGPT</a>. These are examples of fairly sophisticated generative systems. But whittled down to basics, they are a means to:</p>
<ul>
<li>(a) concisely represent patterns in data, in a way that …</li>
<li>(b) they can <em>generate</em> later what they have “seen”.</li>
</ul>
<p>A bit like an artist who witnesses a scenery and later recreates it on canvas using her memory; her memory acting as a generative model here.</p>
<p>In this post, I will try to illustrate this mechanism using a specific generative model: the <strong>Gaussian Mixture Model (GMM)</strong>. We will use it to capture patterns <em>in images</em>. Pixels will be our data, and patterns are how they are “lumped” together. Of course, this lumping is what humans perceive as the image itself. Effectively then, much like our artist, we will use a generative model to “see” an image and then have it reproduce it later. <strong>Think of this as a rudimentary, mostly visual, tutorial on GMMs</strong>, where we focus on their representational capability. Or an article where I mostly ramble but touch upon GMMs, use of probabilities, all the w...
</description>
<pubDate>Fri, 22 Apr 2022 20:00:00 -0700</pubDate>
<guid>
https://blog.quipu-strands.com/fun_with_GMMs
</guid>
</item>
<item>
<title>Hello New Blog!</title>
<link>
https://blog.quipu-strands.com/hello-new-blog
</link>
<description>
<script type="text/x-mathjax-config">
MathJax.Hub.Config({
"HTML-CSS": { scale: 100, linebreaks: { automatic: true } },
SVG: { linebreaks: { automatic:true } },
displayAlign: "center" });
</script>
<script type="text/javascript" async="" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML">
</script>
<p>Moving to a new place can be hectic and tiresome. I am moving my blog, from <a href="http://quipu-strands.blogspot.com/">here</a>, and it’s none of those.<sup id="fnref:1" role="doc-noteref"><a href="#fn:1" class="footnote" rel="footnote">1</a></sup> /s</p>
<p>I tend towards writing technical posts when I tend towards writing at all these days, and <a href="https://www.blogger.com">blogger</a> doesn’t give me the presentation options I need. So, for now, its GitHub pages, but with my own domain. That way, if I decide to move again, my (almost non-existent) readers won’t be sent scrambling to find my (almost non-existent) content.</p>
<p>The old blog was titled “Random Thoughts”. I wanted something different and bit more original this time, so I Googled “Not So Random Thoughts”. Obviously.</p>
<p>So many hits it isn’t even funny. So many, that you couldn’t squint and ignore. And that is exactly why you are stuck with “A Not So Primordial Soup”; which, by the way, does a good job of telling you that this is going to be a mixed bag of the deep and the frivolous.</p>
<p>Just for the record, “The Psionic Poodle” was on the list. Since that isn’t the title, joy to us, things could have been worse.</p>
<p>I have been asked about the domain name “quipu strands” (by the 3 and a 1/2 readers I have) . These were a device used by the Incas to...
</description>
<pubDate>Wed, 26 Apr 2017 08:01:36 -0700</pubDate>
<guid>
https://blog.quipu-strands.com/hello-new-blog
</guid>
</item>
</channel>
</rss>