-
Notifications
You must be signed in to change notification settings - Fork 3
Expand file tree
/
Copy pathspeakql.html
More file actions
115 lines (114 loc) · 6.27 KB
/
speakql.html
File metadata and controls
115 lines (114 loc) · 6.27 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.1//EN"
"http://www.w3.org/TR/xhtml11/DTD/xhtml11.dtd">
<html xmlns="http://www.w3.org/1999/xhtml" xml:lang="en">
<head>
<meta name="generator" content="jemdoc, see http://jemdoc.jaboc.net/" />
<meta http-equiv="Content-Type" content="text/html;charset=utf-8" />
<link rel="stylesheet" href="jemdoc.css" type="text/css" />
<title>ADA Lab @ UCSD</title>
</head>
<body>
<table summary="Table for page layout." id="tlayout">
<tr valign="top">
<td id="layout-menu">
<div class="menu-item"><a href="index.html">Home</a></div>
<div class="menu-item"><a href="index.html#members">Members</a></div>
<div class="menu-item"><a href="publications.html">Publications</a></div>
<div class="menu-item"><a href="news.html">News</a></div>
<div class="menu-item"><a href="impact.html">Impact</a></div>
<div class="menu-item"><a href="blog.html">Blog/Misc.</a></div>
<div class="menu-item"><a href="projects.html"><br /> Active Projects</a></div>
<div class="menu-item"><a href="cerebro.html">Cerebro</a></div>
<div class="menu-category"><br /> Past Projects</div>
<div class="menu-item"><a href="sortinghat.html">SortingHat</a></div>
<div class="menu-item"><a href="speakql.html" class="current">SpeakQL</a></div>
<div class="menu-item"><a href="krypton.html">Krypton</a></div>
<div class="menu-item"><a href="vista.html">Vista</a></div>
<div class="menu-item"><a href="panorama.html">Panorama</a></div>
<div class="menu-item"><a href="morpheus.html">Morpheus</a></div>
<div class="menu-item"><a href="hamlet.html">Hamlet</a></div>
<div class="menu-item"><a href="nimbus.html">Nimbus</a></div>
<div class="menu-item"><a href="slab.html">SLAB</a></div>
<div class="menu-item"><a href="orion.html">Orion</a></div>
<div class="menu-item"><a href="http://i.stanford.edu/hazy/victor/columbus/">Columbus</a></div>
<div class="menu-item"><a href="http://i.stanford.edu/hazy/victor/bismarck/">Bismarck</a></div>
<div class="menu-item"><a href="http://i.stanford.edu/hazy/staccato/">Staccato</a></div>
</td>
<td id="layout-content">
<div id="toptitle">
<h1>ADA Lab @ UCSD</h1>
</div>
<table class="imgtable"><tr><td>
<img src="images/speakql.jpg" alt="" height="80px" /> </td>
<td align="left"><h2>Project SpeakQL</h2>
</td></tr></table>
<h3>Overview</h3>
<p>Natural language and touch-based interfaces are making data querying significantly easier. But typed SQL remains the gold standard for query sophistication although it is painful in querying environments that are touch-oriented (e.g., iPad or iPhone) and essentially impossible in speech-driven environments (e.g., Amazon Echo). Recent advancements in automatic speech recognition (ASR) raise the tantalizing possibility of bridging this gap by enabling <i>spoken queries</i> over structured data.
</p>
<p>In this project, we envision and prototype a series of new spoken data querying systems. Going beyond the current capability of personal digital assistants such as Alexa in answering simple natural language queries over well-curated in-house knowledge base schemas, we aim to enable more sophisticated spoken queries over arbitrary application database schemas.
</p>
<p>Our first and current focus is on designing and implementing a new speech-driven query interface and system for a useful subset of regular SQL. Our goal is near-perfect accuracy and near-real-time latency for transcribing spoken SQL queries. Our plan to achieve this goal is by synthesizing and innovating upon ideas from ASR, natural language processing (NLP), information retrieval, database systems, and HCI to devise a modular end-to-end system architecture that combines new automated algorithms with user interactions.
</p>
<h3>Downloads (Paper, Code, Data, etc.)</h3>
<ul>
<li><p>Database-Aware ASR Error Correction for Speech-to-SQL Parsing<br />
Yutong Shao, Arun Kumar, and Ndapandula Nakashole<br />
IEEE ICASSP 2023 | <a href="papers/2023_SpeakQL_ICASSP.pdf" target=“blank”>Paper PDF</a>
</p>
</li>
</ul>
<ul>
<li><p>Design and Evaluation of an SQL-Based Dialect for Spoken Querying<br />
Kyle Luoma and Arun Kumar<br />
Under Submission | <a href="papers/TR_2023_SpeakQL_Dialect.pdf" target=“blank”>TechReport</a>
</p>
</li>
</ul>
<ul>
<li><p>Structured Data Representation in Natural Language Interfaces<br />
Yutong Shao, Arun Kumar, and Ndapandula Nakashole<br />
IEEE Data Engineering Bulletin 2022 (Invited) | <a href="papers/2022_SpeakQL_DataEngBulletin.pdf" target=“blank”>Paper PDF</a>
</p>
</li>
</ul>
<ul>
<li><p>SpeakQL: Towards Speech-driven Multimodal Querying of Structured Data<br />
Vraj Shah, Side Li, Arun Kumar, and Lawrence Saul<br />
ACM SIGMOD 2020 | <a href="papers/2020_SpeakQL_SIGMOD.pdf" target=“blank”>Paper PDF</a> and <a href="papers/2020_SpeakQL_SIGMOD.txt" target=“blank”>BibTeX</a> |
<a href="papers/TR_2020_SpeakQL.pdf" target=“blank”>TechReport</a> |
<a href="https://adalabucsd.github.io/research-blog/research/2020/06/14/speakql.html" target=“blank”>Blog post</a> |
<a href="https://drive.google.com/drive/folders/1tSxUTu2A7qy8fPtB81RnwkyakgykZ3iw?usp=sharing" target=“blank”>Dataset on Drive</a>
</p>
</li>
</ul>
<ul>
<li><p>Demonstration of SpeakQL: Speech-driven Multimodal Querying of Structured Data<br />
Vraj Shah, Side Li, Kevin Yang, Arun Kumar, and Lawrence Saul<br />
ACM SIGMOD 2019 Demo | <a href="papers/2019_SpeakQL_SIGMOD.pdf" target=“blank”>Paper PDF</a> and <a href="papers/2019_SpeakQL_SIGMOD.txt" target=“blank”>BibTeX</a> | <a href="https://vimeo.com/295693078" target=“blank”>Video</a>
</p>
</li>
</ul>
<ul>
<li><p>SpeakQL: Towards Speech-driven Multi-modal Querying<br />
Dharmil Chandarana, Vraj Shah, Arun Kumar, and Lawrence Saul<br />
ACM SIGMOD 2017 HILDA Workshop |
<a href="papers/2017_SpeakQL_HILDA.pdf" target=“blank”>Paper PDF</a> and <a href="papers/2017_SpeakQL_SIGMOD.txt" target=“blank”>BibTeX</a>
</p>
</li>
</ul>
<h3>Student Contact</h3>
<p>Kyle Luoma: kluoma [at] ucsd [dot] edu<br />
</p>
<h3>Acknowledgments</h3>
<p>This project is funded in part by the NSF under award IIS-1816701.
</p>
<div id="footer">
<div id="footer-text">
Page generated 2024-07-03 22:00:56 PDT, by <a href="https://github.com/wsshin/jemdoc_mathjax" target="blank">jemdoc+MathJax</a>.
</div>
</div>
</td>
</tr>
</table>
</body>
</html>