-
Notifications
You must be signed in to change notification settings - Fork 2
/
01-practice.qmd
248 lines (149 loc) · 9.99 KB
/
01-practice.qmd
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
---
title: "Frequency List: Practice"
author:
- name:
given: "Gede Primahadi Wijaya"
family: "Rajeg"
url: https://www.ling-phil.ox.ac.uk/people/gede-rajeg
orcid: 0000-0002-2047-8621
affiliation:
- 'University of Oxford / <a href="https://www.cirhss.org/" target="_blank" style="color:DodgerBlue;">CIRHSS</a> & <a href="https://github.com/complexico" target="_blank" style="color:DodgerBlue;">CompLexico</a>, Udayana University'
date: 2024-07-20
date-modified: now
format:
html:
toc: true
toc-location: left
number-sections: true
number-depth: 3
editor: visual
bibliography: references.bib
csl: "https://raw.githubusercontent.com/citation-style-language/styles/master/unified-style-sheet-for-linguistics.csl"
---
## Materials {-}
- source files for all materials:
- <https://github.com/complexico/dipscorling2024>
- pdf version as a handout [here](https://github.com/complexico/dipscorling2024/blob/main/01-practice.pdf)
- How to cite these materials:
> Rajeg, Gede Primahadi Wijaya. 2024. Materials for the *Diponegoro Summer Course in Corpus Linguistics* (*DipSCORLING 2024*) (22 - 27 July 2024). R Quarto. Zenodo. [https://doi.org/10.5281/zenodo.12793922](https://doi.org/10.5281/zenodo.12793922). (22 July, 2024).
## Practice 1: Words beginning with certain strings
For this (and the remainder of the) practice, we will use the `Brown Family (CLAWS + TreeTagger tags)` corpus.
We will compare the results of retrieving words starting with certain strings/character using two approaches:
a. Retrieving all words then filter
b. Directly using BASIC's `starting with` approach
Pay attention to the results. Why do you think they differ?
- IMPLICATION 1: some limitation of SE regarding their result outputs.
- IMPLICATION 2: important for our aim to target/generate specific list with certain criteria using more targeted feature.
### First approach: The basic, find all approach then filtering
1. In the `BASIC` tab (@fig-basic), select: **words** \> **all** \> [**GO**]{style="color:firebrick"}
```{r}
#| label: fig-basic
#| fig-cap: "Wordlist `BASIC` interface searching for all words"
#| echo: false
#| out-width: 7in
#| out-height: 4in
# quarto render 01-practice.qmd --to pdf
knitr::include_graphics("img/04-practice-00-basic.png")
```
2. On the output page of the Wordlist (see @fig-filter), click on the `Filter` feature (i.e., the up-side-down triangle lines, to the right of the "eye-like" symbol on the upper right corner).
3. Select the option `Starting with` as shown in @fig-filter below.
```{r}
#| label: fig-filter
#| fig-cap: "`Filter` option in the Wordlist output, choosing the `Starting with` condition."
#| echo: false
#| out-width: 7in
#| out-height: 4in
knitr::include_graphics("img/04-practice-01-filter.png")
```
4. Type in "kn" in the field
5. Press `Enter`; the result is shown in @fig-kn-result
```{r}
#| label: fig-kn-result
#| fig-cap: "Output of filtering words starting with `kn` in the all word list."
#| echo: false
#| out-width: 7in
#| out-height: 4in
knitr::include_graphics("img/04-practice-02-filter-results.png")
```
You will get five items.
### Second approach (your turn): The basic, `Starting with` feature
1. In the `BASIC` tab, instead of using **words** \> **all**, use **words** \> **starting with**
2. Then, type `kn` then hit [**GO**]{style="color:firebrick"}
3. Compare the current results with that in @fig-kn-result. How many do you get? (Answer key (you need to LOGIN 😒): <https://ske.li/15i>)
3.1. Why do they differ?
## Practice 2: Words of certain word-class/part-of-speech containing certain affixes
We are still in the `BASIC` tab. We will explore the productivity (the number of different items/type frequency) of English adjectives containing suffixes. We focus on suffixes meaning 'having a resemblance of', particularly comparing -*esque* (which has a more specialised meaning of 'in the style of \~') and -*ish* [see @bauer2022, 55].
### Requirement
- IMPORTANT: This involves results from two searches: one for each suffix
- Later, in the output interface, look at the upper left corner to find basic quantitative information:
- the number of items (i.e., the type frequency)
- the total frequency of these items (i.e., the token frequency especially of adjectives having these suffixes)
### Task
- Conceptual aspect:
- try to intuite which suffix would be more productive, in terms of their type and token frequency, in the corpus we use (later check this intuition with the results).
- Operationalisation:
- How would you devise a targeted search using just the `BASIC` feature to retrieve **only adjectives** ending with these suffixes? 🤔
- REMEMBER: you need to run two searches for each suffix
- Results:
- How many items are there for -*esque* and what is the total frequency of these -*esque* adjectives? (Answer key: <https://ske.li/15k>)
- How many items are there for -*ish* and what is the total frequency of these -*ish* adjectives? (Answer key: <https://ske.li/15l>)
- Which suffix is more productive (in terms of the total number of items) in this `Brown Family (CLAWS + TreeTagger tags)` corpus? Is your intuition supported by the data?
## Practice 3: *windshield* vs. *windscreen*
We are now in the `ADVANCED` tab of Wordlist.
### Task
- Conceptual aspect:
- these two words refer roughly to the same objects (see Google Image results for [windscreen](https://www.google.com/search?sca_esv=3b07b09e3325eaa7&q=windscreen&udm=2&fbs=AEQNm0DPvcmG_nCbmwtBO9j6YBzM68ZanC7g01Skprhw5JoufVCiMv-hxC44jt6JduRQysBab-bgQXjPraaWFXMvOy8Kr1OAG3K-aj3De4zf3-LxKoaBOHENVo1k3FsU_7VdG-lbBdYQBVAakrHrBWGS-k8Zm9pRqIJKY0zb_8Yc1yb2kIc_SnnCknRRcFFT0aENEckMuT8Kk1K5cHbClUhGfX7kzn-27g&sa=X&sqi=2&ved=2ahUKEwjGlKvCsbqHAxWNR2wGHbcgA0kQtKgLegQIDBAB&biw=1440&bih=754&dpr=2) and [windshield](https://www.google.com/search?q=windshield&sca_esv=3b07b09e3325eaa7&source=hp&biw=1440&bih=754&ei=XS2eZovaFIiOseMP7c6iwAg&iflsig=AL9hbdgAAAAAZp47bX1dqN_ekER2AtQRPVd_8Y1z62jg&ved=0ahUKEwjL_NW9sbqHAxUIR2wGHW2nCIgQ4dUDCA8&uact=5&oq=windshield&gs_lp=EgNpbWciCndpbmRzaGllbGQyCBAAGIAEGLEDMgUQABiABDIFEAAYgAQyBRAAGIAEMgUQABiABDIFEAAYgAQyBRAAGIAEMgUQABiABDIFEAAYgAQyBRAAGIAESOcTUABYmhJwAXgAkAEAmAHWAaABlQeqAQYxMC4wLjG4AQPIAQD4AQGKAgtnd3Mtd2l6LWltZ5gCDKACwgfCAg4QABiABBixAxiDARiKBcICBBAAGAOYAwCSBwYxMS4wLjGgB84z&sclient=img&udm=2))
- are they different in terms of their frequency?
### Operationalisation:
- try to check their frequency in ALL Brown Family first (think about how you would devise the search)
- then, check their frequency in the combined region corpus.
- American: BROWN, FROWN, AE06 (AmE)
- British: FLOB, BLOB, LOB, BE06 (BrE)
- Do you learn a pattern of use of these two words from their frequency searches? Let's discuss!
### Results
- Overall frequency in the WHOLE BROWN FAMILY of
- *windshield* (answer key: <https://ske.li/15n>)
- *windscreen* (answer key: <https://ske.li/15o>)
- Frequency by the combined sub-corpus
- *windsheild* in BROWN, FROWN, AE06 (AmE) (answer key: <https://ske.li/15p>)
- *windsheild* in FLOB, BLOB, LOB, BE06 (BrE) (answer key: <https://ske.li/15r>)
- *windscreen* in BROWN, FROWN, AE06 (AmE) (answer key: <https://ske.li/15t>)
- *windscreen* in FLOB, BLOB, LOB, BE06 (BrE) (answer key: <https://ske.li/15s>)
- Do you learn a pattern of use of these two words from their frequency searches? Let's discuss!
This practice is inspired from Stefanowitsch [-@stefanowitsch2020].
## Practice 4: Frequency of lexical-verb tags in American (AmE) vs. British (BrE) English for the Press: Editorial (AmE & BrE) vs. Press: Reportage (AmE & BrE)
### Requirements
- We need to run four searches (and hence download four search results):
- lexical verb tags for Editorial in AmE
- lexical verb tags for Reportage in BrE
- lexical verb tags for Editorial in BrE
- lexical verb tags for Reportage in AmE
### Task
- Conceptual aspect:
- Do these sub-genres differ in the use of certain classes of lexical verbs across the two English varieties?
- We can look at certain tag, for instance the simple past tag
- Operationalisation:
- Advanced
- Select `tags` (Layer 1)
- `matching regex` (Layer 2)
- click the `TAGS` to reveal the pattern to get tag for "lexical verb" (i.e., the `VV.*`)
- Text types^?^ select:
- doc.genre: Press \> Editorial
- doc.region: American
- Hit [**GO**]{style="color:firebrick"}
- Results:
- NOTE: it takes a while, even for this small, less than 10 million tokens BROWN Family.
- In the `View options`, check the Frequency per million words to see the relative frequency
- Answer key: <https://ske.li/15m>
- There are several list of verb tags, focus on comparing a given tag between Editorial and Reportage (within variety) and between the same sub-genre across variety.
- For more insightful analyses, we need to further process these datasets in Excel or in statistical programming language such as R.
- Most of the time, corpus tools like Sketch Engine or AntConc are just part of the means (e.g., providing raw data) to an end. For example, we might want to directly visualise the distribution (i.e., relative frequency) of selected tags by genres between varieties as shown in @fig-lexverb-tag. This requires processing the output of corpus tool further that these two corpus tools cannot do. The graph in @fig-lexverb-tag is produced in R ([code file](https://github.com/complexico/dipscorling2024/blob/main/01-freqlist-code.R)) based on the [data in the repository here](https://github.com/complexico/dipscorling2024/tree/main/results) (see the .csv files starting with `lex-verb-...`).
```{r}
#| label: fig-lexverb-tag
#| fig-cap: "Relative frequency of selected verb tags with absolute frequency greater than 1,000"
#| echo: false
#| out-width: 7in
#| out-height: 3in
knitr::include_graphics("results/fig-01-lexical-verb-tags-by-variety.png")
```
## Reference(s)