From 07595d7dd3a07114ab0f165830b04977c2c9e5a0 Mon Sep 17 00:00:00 2001 From: zackbatist Date: Fri, 21 Feb 2025 16:08:21 -0500 Subject: [PATCH] Built site for gh-pages --- .nojekyll | 2 +- CITF-Postdoc.bib | 53 ++- ethics-protocol.pdf | Bin 51465 -> 51463 bytes listings.json | 17 +- notes.html | 8 +- notes/methodology-notes.html | 230 +++++++--- notes/potential-cases.html | 2 +- posts.html | 29 +- posts.xml | 26 +- posts/weeknotes-2025-W08.html | 825 ++++++++++++++++++++++++++++++++++ research-protocol.pdf | Bin 70381 -> 70377 bytes search.json | 656 ++++++++++++++------------- sitemap.xml | 84 ++-- 13 files changed, 1483 insertions(+), 449 deletions(-) create mode 100644 posts/weeknotes-2025-W08.html diff --git a/.nojekyll b/.nojekyll index 0cc8700..fd79d29 100644 --- a/.nojekyll +++ b/.nojekyll @@ -1 +1 @@ -3b0ecf03 \ No newline at end of file +550cb62e \ No newline at end of file diff --git a/CITF-Postdoc.bib b/CITF-Postdoc.bib index dff42b6..227ba2b 100644 --- a/CITF-Postdoc.bib +++ b/CITF-Postdoc.bib @@ -9697,6 +9697,14 @@ @book{charmaz2006 file = {/Users/zackbatist/Library/CloudStorage/Dropbox/zotero/Charmaz 2006 - Constructing Grounded Theory.pdf} } +@misc{charmaz2010, + title = {Grounded Theory: Objectivist and Constructivist Methods. {{Dans W}}. {{Luttrell}} ({{Éd}}.), {{Qualitative}} Educational Research: Readings in Reflexive Methodology and Transformative Practice (Pp. 183-207)}, + shorttitle = {Grounded Theory}, + author = {Charmaz, K.}, + date = {2010}, + organization = {New York, NY: Routledge} +} + @book{charmaz2014, title = {Constructing {{Grounded Theory}}}, author = {Charmaz, Kathy}, @@ -22808,7 +22816,7 @@ @article{gubrium1998 urldate = {2025-02-19}, abstract = {While the sociological analysis of personal stories is becoming more sophisticated, it would benefit from a refined appreciation for narrative practice. The theme of this article is that the coherence of stories and the experiences they convey are reflexively related to the manifold activities and the increasingly diverse conditions of storytelling. Drawing on ethnographically assembled story material, we offer an empirically sensitizing vocabulary to illustrate how practice constitutes coherence through the interplay of narrative composition and the local conditions of storytelling. The vocabulary high-lights the growing need to think of personal stories and their coherence as an active ensemble of narrative practice, now “owned” as much by the diverse auspices of story-telling as by the storyteller (Alasuutari 1997).}, langid = {english}, - file = {/Users/zackbatist/Zotero/storage/JPP43SDP/Gubrium and Holstein 1998 - Narrative Practice and the Coherence of Personal Stories.pdf} + file = {/Users/zackbatist/Library/CloudStorage/Dropbox/zotero/Gubrium and Holstein 1998 - Narrative Practice and the Coherence of Personal Stories.pdf} } @article{guerra-doce2015, @@ -25065,6 +25073,15 @@ @thesis{henton2011 file = {/Users/zackbatist/Library/CloudStorage/Dropbox/zotero/Henton 2011 - Herd management and the social role of herding at Neolithic Çatalh-Hoy-Huk.pdf} } +@article{henwood2003, + title = {Grounded Theory in Psychological Research.}, + author = {Henwood, Karen and Pidgeon, Nick}, + date = {2003}, + publisher = {American Psychological Association}, + url = {https://psycnet.apa.org/record/2003-04376-008}, + urldate = {2025-02-21} +} + @article{hernan2009, title = {Epidemiology, {{Data Sharing}}, and the {{Challenge}} of {{Scientific Replication}}}, author = {Hernán, Miguel A. and Wilcox, Allen J.}, @@ -43563,7 +43580,7 @@ @article{ragin1998a @article{ragin1999, title = {The Distinctiveness of Case-Oriented Research.}, - author = {Ragin, C C}, + author = {Ragin, Charles C.}, date = {1999-12}, journaltitle = {Health Services Research}, shortjournal = {Health Serv Res}, @@ -49482,6 +49499,24 @@ @article{sosna2013 file = {/Users/zackbatist/Library/CloudStorage/Dropbox/zotero/Sosna et al 2013 - Burials and graphs.pdf} } +@article{souleles2024, + title = {How to {{Analyze Your Data}} without {{Lying}} about {{God}}: 49 {{Sermons}} and a {{Reflection}} on the {{Systematic Analysis}} of {{Ethnographic Data}}}, + shorttitle = {How to {{Analyze Your Data}} without {{Lying}} about {{God}}}, + author = {Souleles, Daniel}, + date = {2024-10}, + journaltitle = {Current Anthropology}, + volume = {65}, + number = {5}, + pages = {767--786}, + publisher = {The University of Chicago Press}, + issn = {0011-3204}, + doi = {10.1086/732690}, + url = {https://www.journals.uchicago.edu/doi/full/10.1086/732690}, + urldate = {2025-02-21}, + abstract = {Despite often heroic commitments to immersion in a given field context, and despite frequent theoretical fluency in analysis, anthropologists rarely show the steps by which they analyze and interpret their data. Because of this, the validity of anthropological analyses is generally difficult to assess, the reinterpretation of existing studies is often a dubious proposition, and the ethical implications of research sampling strategies are often impossible to know. This essay makes use of a corpus of sermon data collected over a summer at a Catholic monastery to illustrate the limitations of sociocultural, ethnographic anthropology as currently practiced and offers several strategies for the interpretation, analysis, and presentation of anthropological data meant to fix these limitations.}, + file = {/Users/zackbatist/Library/CloudStorage/Dropbox/zotero/Souleles 2024 - How to Analyze Your Data without Lying about God 49 Sermons and a Reflection on the Systematic Anal.pdf} +} + @article{souvatzi2007, title = {Social Complexity Is Not the Same as Hierarchy}, author = {Souvatzi, Stella}, @@ -52692,6 +52727,20 @@ @article{urban2014 file = {/Users/zackbatist/Library/CloudStorage/Dropbox/zotero/Urban et al 2014 - Mediating connections through materiality.pdf} } +@book{urquhart2013, + title = {Grounded {{Theory}} for {{Qualitative Research}}: {{A Practical Guide}}}, + shorttitle = {Grounded {{Theory}} for {{Qualitative Research}}}, + author = {Urquhart, Cathy}, + date = {2013}, + publisher = {SAGE Publications, Ltd}, + location = {1 Oliver's Yard,~55 City Road~London~EC1Y 1SP}, + doi = {10.4135/9781526402196}, + url = {https://methods.sagepub.com/book/grounded-theory-for-qualitative-research}, + urldate = {2025-02-21}, + isbn = {978-1-84787-054-4 978-1-5264-0219-6}, + file = {/Users/zackbatist/Library/CloudStorage/Dropbox/zotero/Urquhart 2013 - Grounded Theory for Qualitative Research A Practical Guide.pdf} +} + @article{ussishkin1971, title = {The" {{Ghassulian}}" Temple in {{Ein Gedi}} and the Origin of the Hoard from {{Nahal Mishmar}}}, author = {Ussishkin, David}, diff --git a/ethics-protocol.pdf b/ethics-protocol.pdf index d3f67b641f0f81ae8a2e9cfb0ac8567b28a57c8f..1f719981c8e8a263c975c94c3340a8aea14ebf22 100644 GIT binary patch delta 7049 zcmai2MNk|HkY&(;!GpUK+y{4e32wpNgWDvy1{vI4f&_;^kl-F9I0W~>-S_XWz5LVO zx~jVmujRe!j!s61Oh%|{RfMcNgmU5yo(07iyTldAWqVRd%@GJKyNurgswspE<)Uwk zT*W5{h4dvh({sev)5-YPtcB;|#$RSkGG~X6$$1i-t23M?Hl5nb;U36GT2 zc2rqxro9d8rN2G;P`H(t99@=AeQL{B*~}w+tS|hN2zLS3FBydgumhoCAc=irl_axG zqf_|ogpb;wqZ=Xh8yoUlqGeQA;WO?OT1v?LfyiZZdCr5$Rn1t_Gu6a@`KQCXuU z%m-jSgPWqb+D|uy?IKl>ALe6!wPCW1>Bh3D(3>Ad`vcivZtL))%d7W>P-c+Fsq)p- zt?WKy{{FRK*{;d4z1uivc2Z7xx#o4wUDxqG>(vYV{rJ!bnl+SHSpdKTbyg;X_`!%| zNuo4-{-#OqGiCwR#pOs*N=*~MjARg)6`Gg%({GwOtlf$_Fx_pL$ z2jd%bb)1)+X`eFBQ)=q+@+re!aM(ECwvV-g_r9-3Hrj_{K3t;TxDmSeb;A0> zm?|nBFY9{l#|LTyJw=7ADSRep6i)S;d&nFgU;P2A-Q&Stx<`%;-F)VBB<`gfm(;iuO${9LG zDG(wF?1)V8p6<-fG6)Lr2a{w)d#i{q*!ey>$LDIdy}>a8(#)Z?Dnbyjm>Ak8PY-Kz zXEfhs6R4_VuLO3)*#mw+6W&PVk_iZRs-i}t>fZf;i=h}JLsftK7i8UP{g)I1sz`RG zpRKpevnSuj%m)0|`T2)DLw9yn^F!TrXSFSv^ci1sx)^TP%vJvH4TD#+Uh{$0_^6Pp zmy6)VuHaf9uhdLVqDjb+BbNw8S|Kh{uDQQ2FCER^T;N+!6*nz4)f?15Zj?k&0pYY3NY;VhB<5_0q2+# z!`cl)+-$kwBu#h}t>Q`OJfaiJW!Q0s!rJ91>DKx1@EjHnVnW(w^!_nas(XSp5)Y(( zL?@F`@$Efjr5fSeiwpoJ)73rwH-$79qVX}j>&gDNQXH>pTKst;kuL?e$2Y6B&o0jV zr!FXKG&ejSN&IW=Kk#>^2S5PzPsiI(Od*z?|)jEvf{AKi6^`Mg75D{9K+F2 zf*kTI+7j3_Tc{}BQ=b}F*i|5a_3%{R%+C}@{iQVa_ssMKYU`85dMgz8)>bICsF<_T z<$f!)YV15PV!A9A^=`W9M&|O(OD|rsE$I8;x{j@^e&Ei~*2%)k)BcoB5QMBVaCz*c z!-M1F5chmhy=}GFUwIbX%PxV{@6eOm7eDM{FDe-USI6(rj@a^_#A>&G#7}PVUd|94 z^^Yr|l!B{+7c1T^`7(G`nGh57PJcU?SGB|Ow^xf$FJAFX{q!p+Qj6| zl{JC?>hitP_W0~P6qz1%4}zQ6PsTKqgSg3-71laTrtO=BB6?ECft0$4lFYS-2b|_Z zo!(_puq;I;uBPj>puek;h?8(cDB_E@iId>bAlXDmippcb7FA>d+Ny6>@>BVi4a=aZ zs}>Q{PQ`@Qd)lA*B48XATTsr1tP#aY7vVs zNf-2P(s#H(s?#om7{96XZGPrPKZ*6;7m2id%($XXY_~Q`|CPVd|3h zJ?F+ICZCUN=_T%-V>xDwJ{Wg^WgYkagk!?|&>^8rcReTDD9$yg%a1Ln?Bt|a_-kJq z?6Mo8^_ef{-?^dIWCZX-6*Va#?g~d*ce^_XR`w@A3&rZ)(6_E+3y*^kMByxh2`8<& zNJAEs5q{$7qEJ(NMS;vrM5)No8P+jL=FNHbzZN@b^g4YfjFdRZvjH!+pZs4p)D8y~ z=+@kP)?T+b*9Dw8_APJGaGyH%e(Hw>o*YYes9+8m-XX_?0O(B1800~gPTn(~_jq+2 zmmg9o%sy2vM>RDz^OI=`D;O9Ix-YGjoD(pNA>l3{G|=4WNV#9@aR0enfBJUh@4maU z^mIMYcomo~cN>FZ67{0MW<&E9u&O$a&-y+-;l{u+lsQlQbC%c1zlc9-fg!&f`Z&=$ zpIAI$M1Agq-e_MEanq#dxm&Mci%JgqPTK;J&o37I*L9(Y{{HGS?`l>o9X;XvZ*@J; zIqix?CcQ6k3x_mGdp$G^BRZ7~)Ov9YeR_jpQNR3kdG>tuKk=Qbhn@J_FEYL-qevM$ z!oKWuNeIkEP^%$k#X|xVh}*363*O^nrZ$Ln$-s8-{hj3mj2-i^y9DlH-(|rOGmY_n z6?IA0-Gqh7aEwQl3^(A$lrf=_+TBG7f0~I|cLWg>bThL=*y;k(TZYm<2PMg9*@Ugt9MQ~7 zOs;7bU$~jc;;fV=oP?zc2(l-(5hv|N@}Uj?cEm%Ef4=Lke3X3lK!Y_HFBznD$=*hw zn4CW~vDg%O`HLuBqD}@Z1enA%BXMb0Sr$@VA=h)F_Y;a0u0nUMsf0n*{30wb#Qo|`W?hGXcdh|lmgs^ zhi{@0=<7`fzjwk6`5WNS@;DPi=N{%3+4A@|yF9~6C!nnL0yd^?%r>rn$`R@D8yYO6 zg&hK~VZ($b30mbUMtH)dh58(b_8 za~wd-2ka^SV{e?Bw&so)H+mff;Af-*d_jNhNEKZf!a z#J03E6^=jdV3WC2Q8B8-f-;RoJY5i~_77EMMI%bZcFwRW{zFw@JX0>Wv7)Ws6GBSt zqDy0b%KiP{%Ff8O>$iNjv$?ac-lFsikT@`(+`eSe*42f*r=J>rPqgw7@{91O)H?jA)Cw9!r zwF>2mkQY%@ZGN#J`mZa1OfN{%*= zgP~%umutmbSAa1~G*l|>!w~jPg=H9mlV_eW z1lbMwa=?e~%{UIB5Bk|=NOt$oHy#iUYg~iu=&(at$CDqVcpNo)^&K^Z;1MC#pT%OB zsM$i_BS{mhXmiH-#X-r=D)bHzHAj1VB?`&hE{F)>=-aGP&_UH@9!9%IO@b>p;)c`(7bVk%>759U%{~&`=Kv)VuN%5);wPAO7o8l@2Zwb?Y zd(7dZo-b@@70ek)5*mr2cvO&gAtDj4@ESz#Dv*1fsJ4140UoGGD!a4^)xAJ~b4{{c z@A$j~|NRmi9l)$wn!gFf7xNVFJQof%7tsQ-JQ*E>z!&tz2dfL9dRj@Uk;cojp8`NCGoQ# z;S0(#>M0pl4nsJd8`cpJ?iFSbeEaZi*r$BfAT{5#`}@I(ugDASdiI3;z|W!jQD1RL zCgtvv?+E(q@qP!9Z&r6k?!@1{Na5{=r&BvHYh%^4AEWre#j~67O}_Z=b%MTcDDrlF zJmKL{iaaQj_Kw_ybnTGmeMdX09}~P>Ojj~yjNEKdn$#Z*1Cww+cW7?zx5{6)UKM_; z9X-2zKKpf^a2M)eth?mvFE)0UFp1g0*w_#foGVAWXf(gCE!m1ypV!`~Zt06{!Z&pmjWI;>ykxGstb=Fy}zVZTn`6j&>PQA|DJc zDn}F8Q3wZunQyh+ul0M-QLLxgVo_lfzsr{L(~FVN4J%m{GOV2`xviOu!{a~! z2$7}rA-#tf9Rw9ljvLMwAm4@1tkny<*}%#IBsQ1&&8m*J&DW2RmsAr>jzOp*m~5wS z1m^#TJbr)YAps&=j*lHjm|1dA8f3%thLmk!`$is==+2|m&R~}^tZPP@9I}y_vRAK7 zg7HoA&Vv&PmOjvAXo4IDj~lfqGMnbjF)S!$efjY!fs=$Q`FWdL?F(t5?_DavjvkmW zTKv;+t!@UocFL-i!k^YMdo?J6JDCIi+323VlGl|BTy3gtLJq>D_?+I8Oy49n9E z^_yGzA@!+D_~m)%mv#M&)mC+7=mAl8Ae7w6fY9}u<-?__PhI)1m3U&y?8g5R?>KGLoY4*D?mjhW@Oi@zDaq2RJIaQ{AbJ0Q=R_CnO=Qw zKwhYg1HXk7Zx3~%BC$r37O7Xfc#%lT+46G@2naInhFWPHRp!6R!S`PZ`w1FStW|< zk9=DA-Q|Dkn43FYvg@s#lZ=6iPgxl<4r3}TKbo2}j+=oVIw1N5tkyfSzj-ft4)PFb zgP5^aModx$U4Dk!SgT5?L@rtFJ66N*?(r>m?`kK$IbP4#x+_Lo%PkyNSb#O?Ge{p- z3zQt0n!+ldPxb^mnDHs>2BUG~2OWrEymV}7SuV{G2c5@ciBe>cek6gwo2;Ojo3Cp>wu| z9>2BypZL6yK<|aQ$Z##2I+b@^L*{j9UKF!q&+?kMLow4gbs1-YBtc{!<0m=9>(_ng z(z6^-gU>vZJEO8uLcc8v-eampC^OZCx(#~&c!;=i6++G8*BkKUM_%&Jmd%G;(cFHw zEv9+UEtkfp!UQWq)}!I_clNt9TvD_e_9%bD-*_tQSyork791S`dB z^qQvqUrcX|e@bJ&m~9n%G*AXY)RmVASIwSEG$RD$fJ>oc8q4k1l__M#&F>3&FEL&p zdpiUG21xc_?ov;e3W{V$ujAuWb!n5AD|q+lOT&)?h*t{2X%wg6jr<i0Mhqkr+BUbxU$$myyzG18rXF3ApIMJac?0n?y%}OOQZ$S z>I8wPDTewmFb2XXr$jVRxKgTaNK4MB7?aNrn`ngk^RhdZh||ZhJZap8sN-Z8e=-xe z!#`nuKt0}RDug#G0%nNIA>ZqdYw9TMm?E6xgV`TNEEM~+QX-ld#A`IgT5RqZ^1Ga@E7P;s^B!DE}&+*l?)S%qioapLg zcKUO1c_)VX86GdHpn+a*&a_jgwgo2>IsgdNEFCr@ZZ5Z_0Yf6@`@vE}rAIO{hO;jr zmbfAX-BFZT62jW@VdP0#@<)XQUdntI6fvg@PKfwyh~LuOUy5cj^UAGZTPiXFo`6j^ zsknF8{s(f2k^vL+b49$-=2DCm3Y|csD37oxFMIM2Owc?|4CTb$$PM-P_U7wrj%2e9W@ajkV4tD;X1Y@pK(`KVH%N0YrZ;_Ns^d z+_h;~(Dcaq)LqlXwZ*<&7_UIBm=EXJ+#FqSp%qzxQF46ovYS|!R9Ota8J;56UjR6- zJC@(6Y9)RS=fOoVjyTOo=yG7c7APbhuKQL@%7nO61xX*c-;zFRdb-$J^VJT5OA{Jh zE?!-lOKYq}*U$KR+8s9Y@(lbED6*Pss)9Cn)i<(HRHCeFE>HrIV-Jit7tc){;Q9M} z6}Y>8wJfDXeT-8w-*-`>;pv`sOVHC)alqghzX@Vo>l0o>zCIm$emZmmhtOW?nuIv= zeui&sgQP1yMK><}p5^i#(GL%@O*3K zcXd2QXv+w)&SR|r20)o!B_(AKRJH4sS}H@FAYTeB4oC?~19yssFqx03ouC6X$@ET= zY0bT@)=F&yDRrVBF#p2x$jMrH{o3oa*4fxblDpCZ-StsP`(iMiYb!E7W9#z1jBX|M zjmkY6sw7#2GaU;57R_u>d~LB(RQx_ckb}@o2O{r$J{)Kr62MS96yeYBqR^(yOXhuK zfY`&N6^b7Q;Vf4=2qeWjk1RSS4{bFUxk<-b3sr?`tArb-a|K#PEOke=TSgUqCK~v{ zVa{kT+qz2Cp6aBrKwn}AEL9A9e9Ymk&umirdDKvxz1sKg^m<}s%=KgbX&1pWf#kKg zlV$qp9tQt?2=Q}-G9jIPV24?l?c&G`1gjqJ#Wa?7y@@{~x}-!PrwN)4F~!;|c%QJEQ9G6)3P~7kA4ae_wbK75* z_k8OlpXZfXR!t3Uo(`T@4moT)x_Otw7{o?~s>im^Qy`wfLn8Z7e9k@XI`I!GeL4>sqw6tv{x zv9`9hvbN;s<>Th#<+8Ho1M`4wY^=G(ME<`BsKAi?f4}MU)2-Yx=0uSfXZX5mEZ^H$ zYzbKQ>v~588++4;!gb$kGpCNm6>4+GO%myS25Gbdk43J)-F~n8Pg$S7axoC0b6u2C zKK5WhIWc)Sk*+vF%5`w?a;8G@g}Ua%^JXBh8Ok#@4TB^{q!a?88DVd~H>A9Lb0eul zDi>^*tkOLRw8L3wo>Z_4wTlWNib_?AG>r~biz57;C8qqD{jhy16S%h3Ig^QKkcg$` zCJ9(9#EvN-QYyeprY7~EBD^d>*D4^Jp+?~T1*l9!#@7I)Y9P&NAaqUvy)>{UrZ8)! z5XiX$EnbSJgF)-r1NT(}wkfj~{w5#2By8Ag1od3FT!SnOoMu8%4TR=NQa+nF8p6vP znl8P~WI+*H=M!^Y?b$|`V&>~sbL?t{ClcnPRw*2uM^8HDD^~LzoyT6v)6^*_n{IcQ i3>{VSAkFi;+bbCmqFB*o=;kmLkQ@(MW0=nCee*ytW!_ zWlB<+za~ZByXhX;p(MQYf6j+l%g+LP+_4Uzxs5!ob~<(L%26Re@)l7W%l(A>FJ9u3 ztR1M%LRjL{tD%=Z*~;Y6IpPX>kzOI5$U{@nFEeP%NUQ~ zb(IbVd%_3McM12yptVoE=X2N6XJb>)sDf!;J;aN(jg_|9 zkI) z$p}Pu(h8;2=b%^29n*P>%D>Hl{#PS3TY%_$#mS2~G%s}VNM zd05g;bY>eE{@YEr$iZ?ml^Sc3u#JA&E!+}3AGTUh*m`=4=NUTr8`lXYrX`6Y4D-^W zgx;Hh;D#WiC60dkI!)2?ckK<{Sm;WK)xZoP{3OzdHR`2}x?bVKcj`NG>jiA~w zl-dc`q$r?tg70Xra3QuL(B1sC?ibCIyE~1qdh0BIRzFESPrsDcJ#5TUjXx1x*|q7t zx$!5%ajozE`@0!?dt|&M@*1r(p+;KW2ud_m0_ z-}Q{%N(>F%7QPS^yJ1u?@m(`P_XEv^G4-M6oK4 zq6y3fbRuU`lq6sum4!qu!q+2cq$e36mGN$R?c3I4f^6DWg8=~Iac!AKYyXGvF}?xM zE|olb(v#AdMC;1j$%3fxujQ(Tm3pkA(`k2yG;w+#y<1YdFJNUl0Id7rN{>DSm!P-HuP-7jwK6ZU5=<{nk)$S_!t~E99C`gKTdf7 zp6+isHXNJs&Khe1NHKT80Z{(Q$x&(J(sEI4qIYnWv3&m9cUEda&U2|bvmoXN8^*hp z^(3m3YCQ}4cnh1s;nLeF@4-PKm4$~f^k}4kLfMFR^$>c~sPmUg1C4lD9Hz2Tt;mv%9j3Vl@YAmLxK{jaqpVmGoZ84$trZ9^37#FpJ;lOE zmb+fv<;&H7uWBa7t5D4QfgC|Ti>qzT$g_x8s#d?^uW=E){&~R`v^yOjiT|_NcK(d# zdNMZfDTu_PsA~MtFO_>l`;>9C*Preg% z?F5?R|o}L^?PdCk_L$I~fhqLAHZenjc zOtLF#_cW-x=W7aVpgoFsywvR^dGGn8ktXiTQwI&yq=3AxR1z-#6#aY`PS*>qeK&Gg zHt# z45dIFrNBGXMe`a(OkNE^^XVuLVphc07+0NAYE}e(X1t^_X6Ye~ik>0r$U150PkQV^ z9Fg}!Mljb2Y%x~mqG^=)NOwRzdns=yhZJ_<2XwvUK9i}#{$bL30*!5GyJX)8wY_Rg z(Ay*2!o-Wc0Z|5#YJEW(=R|bTgzyC+QMT`5{*@38KQ$U>0IFAmonz-UC+|D5!H7XS{}lrHpmpI6`Ug;!ZUuFR|q*c?%mxO}tf~@0(Y&6y)N-tw@E}gjK)) zghf5>=o&PzTh{jUc|lu>&cAsJ9*TrvC2uMR6cQ-)BEJmgswF9d3ljMpUcAqCtyx&7 zmXN8jinM5k-58PIH209{D14aCks(wkw1*!;tbYq}_a`j)zJeRQ<9idepPn`7pUCz+ z&{W$=3iOt}XN|r(726ztHL&x>JNZ$YpTpzv0s>@SwO1sOTXku-Fy4MZNtX+{{}=X4 zLm*VvCte7D^NJT_YhuD|d={ThY-Wl1H%n%d4xr4M#fA&{v6a1tAfWf0p1)X+t_+Is zOeIv1AIPFZ!~e?luFa-j<;ZV2v6(It4rwxk!+hgc3Vbi@m5K#tSNCtu26s&+Ta&J8 zAoIn(-G;h#%2T(!vcH1}B|Ds(^~um3p~UkOZj>qM)gwZ71jwTxI(^&`ByU-v$Opi*By6qZ+& z*Ar%$5@s-plqzBsEN1wMk>bQtL{wOOcx61oZxT5^#IH5;VM9=kiR~sL>hVw1UA><3 zkMJ%ji2SCMm%yO|v==RTGcK}_U#WBXOUZvyqCy5OZ0-d>(5S48+SF&Uy_;|isKHhd zVQw?S!7}T3Mvt{j*jdkI-*??oAOukRgaKHUZH+uBi+MOIv#b86>w~`p?@Vwv-$u;O zwO)(2Qi6_Rx2kM}N$!p>xfaaK72P7g&(hEu3qQX|uJ%Vk{u>DgQx2$@kkEfMQP`}< zKWaFBamR~v<<7XU!N5@|1J`<8?k-wsbNs7@SkLzy_lXQNuYYY6gk6b%Hor)0{OqMX z&vtvmdQ(~(pz z4|~phw`R&chZ^$K+}MgJTpeTy#q8%f2q~=*)E_?-H^-{xFvh$qK zLe!R-&xU&tTg=}jB3PH`mDuP-BnQ*%u9&eWuMb-|2Gtw+$8(r~OrG3Gg`DD?VrpMT z>&dmMc#m=PS*rU?--etYqC+5@St?pOxzDB|!&M`4g%R-CP9<%dRqR;05o-Eg^$tIb_n;1oloGNfyb?oMZ_I?=`zH zTi>Fnv}qHhajczIxH;4kev*)W%nRiZL!TN&TVHX({`h(^AT<5kc^O4fx__I$4L>%`27 zL-K|eTk^j|96LeK|A1}_wZOCc??N57-0n_C?$}T)#a_Myu+7p4`P~v3sk+6`cawxH< z?9Q@V)}Y@^l%^+(%f09DGi6)kv^`dV(2-JXlFsq#K{Mm&&H?zlxPmim)-vsV@wA+P z)EYW;_v8N&^T3m za`>G}+4s_qB_U;iHnaZUZPE)zhq>zbf5Sw`9zirX=oi}dQ<05; zGJ_FwsDr*GDtbkdoXC6Y9Q0B)d>y;ecu4Z7B+XP~lBt+wEYaCsLW$lG**PHba;QXNIE0YK}Vl{6Ilr zK5}9Eh9dnhO_ubn zO|n)(G3ps&WfhK9wqnKjAk?;iAc$Uhk&H&=ut0FeP#7m86*;M>#2Z63%ngpkMC+HD z>_gj%;RskJ;KKVr?W<}2x+b4_heVV)EpBjb6ehHcw~Q(&=w@5}gFN{~7(vjB(SbNx zNhtUD56~i5hz^J)m~OtKA5r3+I6bYlfczJMb6WS+wdT>6%d>p#zV%fYXM zL4{Aijl5a4yZ*`M8FTrFTzaSbPz~{?6PG@!Q1~NArl`OvoELpw1YG0Qu zRx$9w0n&=VMzFhwHKEH=25xoAI;0YPd?XIDW(2KALN2USCzJ(ux282wKwAn4k)5@WMN%XoXruSyh^BTDuoaiw~?S7 zT!MRt7T%;=QR2IAG%?Fh2cPiAt?D|pEN*Rhb+SYG%NtX?cXy%k&QuZ4vc6@zudn;y z=_T|r%a%ZJ5~d+d=Quij>bwl4Pd(hbn-8lZA#UHTDL$&OcISEygVA_AQAefw8EN)@ z<;|QjwPVacb*Ul#+hR;ZFr#?uE`)JvcGqrggBgzscn_jx+_CzC5!hw1B#4vdM#PX(#o+- zn<&oodteL1rDS|L@y+Bu8#MU+OKYJ@3a;)X-dvMC+R%PT7K%M0Dxi~1)QLPPD; z#k81fvzXNe${IXg6$%VRKpq=z3VaiiGsFc;XOhd7kUG?i>C5+?PqdA*If+<=lUJqn zj{zVq_p8>&b%I20+!9ctIfDvfl`o}1?2nkBGw=#xi|FE~Aw#PAq)$^U{ncD+AM;os z6>b=NG*V=AFXfQkmQc?+Ei`?n%<4)=a_T*!hgWfJm&>7)HfqOS3#x?Qy(?u)BQMK?$Zg$; z`~?j%*jPttJgt#vb5-U%mMtYCS-ctWXCe=OaEiHcpzkWV3-oF~53LauV9%bMME26g zdO|y9OXtH-?qTa(^WQ1%PV)Rk3P;ExR@#isHC8r#uP62zzJ45r-b*IwIcKW`c<`S& zdCs%EaHG4aICmSp*R;p@GuV1Ldy5aRJLqo;($id8Wf-%MVy&THm_()y!>6wKPu!;O zO~-&=uGZ{j+BdjnE`<=&g#;nk{TD%@nNZW~ffH`#zxi&SMP!h(CDZtp{2GQ8H7XTXn7LA3B8n3gQ&>!aG4q--cWi@VQVIzB?H;j$S1uYuWs}N0 zOET|RkRnhPE)ZkvIRc93{otCO2!#^kv5Q{QJb6A4t6*a+ybiiPwq7KE7lgu0QIq)A zJKdlL-kr$$vFuZ$JMpw{z1k{!hwBHaset%i&T3boVit=Ob!@o|-k9K{qcV(+7zNOf z>dO1{ggL)CbyLM`z@H*^2~v;aTec>7(YvjgS5oA8B)$ywO^3)qt^l;kEM%0 z3(K55N3~TX15vhpmGDW4!;e(c6(h#czmzV&hggHF)dybW5&26J1^Lk%ahDbf5oo24 zV7W+SK7!{;E~MY+LGIPCuA#aFd>gc6VXjfP=MUp2FvfR!bC^r9ZRg)TMM@i8UcM3e zDE-0qvl8k69p(PR$=tasQuxBL={VwRXn0S6m=XQYRwP$Qa27-ME}ZFF$Ff*ixSt{`XFNbBhIkJe%>(?;rY zC|Dpe@t1r7`JJfgbAadm*42!XCUYK1`O4@eTXvuy#Vy5K5y?*#^n=dhf_ z0)VjL2b^layqHdHb&~%0k!hv+h9j<$QTF)6WFdd5v?8@C?AT#cEFH3n}ssSZ;#6u_7*=pw6XQh-4 z)?dem!1#i*_@ePzJ;?_7jUH%>9+28n($caKHzg(0O9@#&z1=k6e>(Q^MpJwUcxQk; zEXt@#<_6ciZ`25R8~*&`?;l*Foshx^kqVusPRC@8rfG_Nw2#cq{c;ccU-}r|`O)Mr zhIz}$QfC$wWvB_IcL7M+#p7FCEX?}|T~x7{t+F_hbrwwGwb{am;51P96ZFD>dO6UYS#pRb+{ zv*io(&kB7zRf%ta8*3GP7PADMc9?zpix?X2TLM(u0O+aXagupPbklgc&j;Gi>V}4N zx?iVL23-=BX+(7*U>B+#+JiGEj18VkF%a;Mp-_x?j&WWX$U zGWmAi>zBCM+RsY`?*Nc1$;X?;-zQpHK8ywLh@zi8W`3e5dIu={V^`elRNCw(!PqlOW3WX3zq94) z1dWoEgg8)2R0JaCC@v=M2$2$%5*4)<5rK#Tg~TC_|7-|A|KAOm=%gA{^nY&>_!oep zDh581j@>-Q#2gcaRS<1LpSBq@uNlp%8P67*1Xi4_T!Qh? zj0FMsRY9w^f25ZK3P=zf7PbZN5zI)B0s{nwFfl<3F{k)?_&DYOv0$3cBm70wVn~T7 z$=~sy3Q-*YV&aq`!tr!`c70qYeM0hd)Sh&#JN*|~>9{88D7onv>*<)s`UC{}XdU`k zH$vDA2<%w|;U)qH4Z-SAwL+^S$epu=M^RC0%|5r^1S5kvz8GyKf94EAWV(&oP!^V5CV`3kv?qTJ<)xp6Q+PDw< rk+f0{`sQ)=sZ7U1vk=-2`i*q@%-EwTiwE19V!{UEva^HrRB`_YJEcf= diff --git a/listings.json b/listings.json index 9655394..9260ca9 100644 --- a/listings.json +++ b/listings.json @@ -1,15 +1,8 @@ [ - { - "listing": "/notes.html", - "items": [ - "/notes/methodology-notes.html", - "/notes/potential-cases.html", - "/notes/maelstrom-readings.html" - ] - }, { "listing": "/posts.html", "items": [ + "/posts/weeknotes-2025-W08.html", "/posts/weeknotes-2025-W07.html", "/posts/weeknotes-2025-W06.html", "/posts/weeknotes-2025-W05.html", @@ -20,5 +13,13 @@ "/posts/2024-12-09-first-team-meeting.html", "/posts/2024-12-09-hello-world.html" ] + }, + { + "listing": "/notes.html", + "items": [ + "/notes/methodology-notes.html", + "/notes/potential-cases.html", + "/notes/maelstrom-readings.html" + ] } ] \ No newline at end of file diff --git a/notes.html b/notes.html index 2848db0..98dd8cb 100644 --- a/notes.html +++ b/notes.html @@ -296,9 +296,9 @@

Notes

- + -Feb 20, 2025 +Feb 21, 2025 Methodology notes @@ -307,9 +307,9 @@

Notes

reading, general thoughts - + -Feb 4, 2025 +Feb 21, 2025 Potential cases diff --git a/notes/methodology-notes.html b/notes/methodology-notes.html index e5ddcdd..79471d1 100644 --- a/notes/methodology-notes.html +++ b/notes/methodology-notes.html @@ -274,7 +274,7 @@

On this page

  • Theoretical sampling
  • @@ -302,15 +302,31 @@

    On this page

  • Coding
  • Memos
  • -
  • Preliminary analyses
  • +
  • Preliminary analyses +
  • The constant comparative method
  • Situational analysis
  • Various other analytical strategies and techniques
  • QDA software and tooling
  • -
  • Writing
  • +
  • Writing +
  • @@ -346,7 +362,7 @@

    Methodology notes

    Modified
    -

    February 20, 2025

    +

    February 21, 2025

    @@ -400,12 +416,12 @@

    Significant Concepts and Frameworks

    Multicase Studies

    These notes describe the features, affordances and limitations of case study research, and articules factors correspoding with variable kinds of case studies.

    -

    I do notice a distinction between two schools of thought, which seem to be spearheaded by Stake and Yin. I generally favour Stake’s flexible approach, and it seems well aligned with other methodological works I’ve been reading (e.g. Abbott 2004; Charles C. Ragin and Becker 1992).

    +

    I do notice a distinction between two schools of thought, which seem to be spearheaded by Stake and Yin. I generally favour Stake’s flexible approach, and it seems well aligned with other methodological works I’ve been reading (e.g. Abbott 2004; Ragin and Becker 1992).

    Stake’s Approach

    In case-study research, cases represent discrete instances of a phenomenon that inform the researcher about it. The cases are not the subjects of inquiry, and instead represent unique sets of circumstances that frame or contextualize the phenomenon of interest (Stake 2006: 4-7).

    Cases usually share common reference to the overall research themes, but exhibit variations that enable a researcher to capture different outlooks or perspectives on matters of common concern. Drawing from multiple cases thus enables comprehensive coverage of a broad topic that no single case may cover on its own (Stake 2006: 23). In other words, cases are contexts that ascribe particular local flavours to the activities I trace, and which I must consider to account fully for the range of motivations, circumstances and affordances that back decisions to perform activities and to implement them in specific ways.

    -

    Moreover, the power of case study research derives from identifying consistencies that relate cases to each other, while simultaneously highlighting how their unique and distinguishing facets contribute to their representativeness of the underlying phenomon. Case study research therefore plays on the tensions that challenge relationships among cases and the phenomenon that they are being called upon to represent (C. C. Ragin 1999: 1139-1140).

    +

    Moreover, the power of case study research derives from identifying consistencies that relate cases to each other, while simultaneously highlighting how their unique and distinguishing facets contribute to their representativeness of the underlying phenomon. Case study research therefore plays on the tensions that challenge relationships among cases and the phenomenon that they are being called upon to represent (Ragin 1999: 1139-1140).

    Stake (2006: 4-6) uses the term quintain1 to describe the group, category or phenomenon that bind together a collection of cases. A quintain is an object, phenomenon or condition to be studied – “a target, not a bull’s eye” (Stake 2006: 6). “The quintain is the arena or holding company or umbrella for the cases we will study” (Stake 2006: 6). The quintain is the starting point for multi-case research.

    According to Stake (2006: 6):

    @@ -457,38 +473,30 @@

    Logistics o

    Preparing to select case study data

    Yin (2014: 72-73) identifies five desired attributes for collecting case study data:

      -
    1. Ask good questions — and interpret answers fairly.
    2. -
    +
  • Ask good questions — and interpret answers fairly.
    • “As you collect case study evidence, you must quickly review the evidence and continually ask yourself why events or perceptions appear as they do.” (73)
    • A good indicator of having asked good questions is mental and emotional exhaustion at the end of each fieldwork day, due to the depletion of “analytic energy” associated with being attention on your toes. (73-74)
    • -
    -
      -
    1. Be a good “listener” not trapped by existing ideologies or preconceptions.
    2. -
    +
  • +
  • Be a good “listener” not trapped by existing ideologies or preconceptions.
    • Sensing through multiple modalities, not just spoken words.
    • Also subtext, as elicited through choices of terms used, mood and affective components. (74)
    • -
    -
      -
    1. Stay adaptive, so that newly encountered situations can be seen as opportunities, not threats.
    2. -
    +
  • +
  • Stay adaptive, so that newly encountered situations can be seen as opportunities, not threats.
    • Remember the original purpose but willing to adapt to unanticipated circumnstances. (74)
    • Emphasize balancing adaptability with rigour, but not with rigidity. (75)
    • -
    -
      -
    1. Have a firm grasp of what is being studied, even when in an exploratory mode.
    2. -
    +
  • +
  • Have a firm grasp of what is being studied, even when in an exploratory mode.
    • Need to do more than merely record data, but interpret information as they are being collected and to know immedately whether there are contradictions or complementary statements to follow-up on. (75-76)
    • -
    -
      -
    1. Avoid biases of being sensitive to contrary evidence, also knowing how to conduct research ethically.
    2. -
    +
  • +
  • Avoid biases of being sensitive to contrary evidence, also knowing how to conduct research ethically.
    • Maintain strong professional competence, including keeping up with related research, ensuring accuracy, striving for credibility, and knowledging and mitigating against bias.
    • -
    +
  • +

    Yin advocates for adoption of case study protocols. He provides an example of a table of contents for case study protocols, which generally comprise four sections:

    1. Overview of the case study
    2. @@ -556,16 +564,16 @@

      Some sca

    Grounded theory

    -

    These notes are largely drawn from Charmaz (2000), which I understand to be a fairly balanced and comprehensive overview of the Glaser / Strauss and Corbin debate, and of the situation of specific methods and techniques in relation to these different stances. I also value Charmaz’s position as someone who subscribes to her constructivist approach.

    +

    These notes are largely drawn from Kathy Charmaz (2000), which I understand to be a fairly balanced and comprehensive overview of the Glaser / Strauss and Corbin debate, and of the situation of specific methods and techniques in relation to these different stances. I also value Charmaz’s position as someone who subscribes to her constructivist approach.

    According to Charmaz(2000: 509):

    Essentially, grounded theory methods consist of systematic inductive guidelines for collecting and analyzing data to build middle-range theoretical frameworks that explain the collected data.

    Charmaz(2000: 511) goes on to situate grounded theory in relation to what was the norm prior to its invention:

    -

    Glaser and Strauss’s (1967) work was revolutionary because it challenged (a) arbitrary divisions between theory and research, (b) views of qualitative research as primarily a precursor to more “rigorous” quantitative methods, (c) claims that the quest for rigor made qualitative research illegitimate, (d) beliefs that qualitative methods are impressionistic and unsystematic, (e) separation of data collection and analysis, and (f) assumptions that qualitative research could produce only descriptive case studies rather than theory development (Charmaz 1995).

    +

    Glaser and Strauss’s (1967) work was revolutionary because it challenged (a) arbitrary divisions between theory and research, (b) views of qualitative research as primarily a precursor to more “rigorous” quantitative methods, (c) claims that the quest for rigor made qualitative research illegitimate, (d) beliefs that qualitative methods are impressionistic and unsystematic, (e) separation of data collection and analysis, and (f) assumptions that qualitative research could produce only descriptive case studies rather than theory development (Kathy Charmaz 1995).

    -

    Prior to Glaser and Strauss (1967), qualitative analysis was taught rather informally — they led the way in providing written guidelines for systematic qualitative data analysis with explicit procedures for data analysis (Charmaz 2000: 512)

    +

    Prior to Glaser and Strauss (1967), qualitative analysis was taught rather informally — they led the way in providing written guidelines for systematic qualitative data analysis with explicit procedures for data analysis (Kathy Charmaz 2000: 512)

    Glaser brought his very positivist assumptions from his work at Columbia, and Strauss’ work in Chicago with Herbert Blumer and Robert Park infused a pragmatic philosophical approach to the study of process, action and meaning that reflects symbolic interactionism.

    Glaser

    @@ -577,10 +585,10 @@

    Glaser

    Corbin and Strauss

    Strauss and Corbin assume an objective reality, aim toward unbiased data collection, propose a series of technical procedures, and espouses verification. However, they are postpositivism because they propose giving voice to their respondents,3 representing them as accurately as possible, discovering and reckoning with how their respodents’ views on reality differ from their own, and reflecting on the research process as one way of knowing.

    -

    Corbin and Strauss (1990) “gained readers but lost the sense of emergence and open-ended character of Strauss’s earlier volume and much of his empirical work. The improved and more accessible second edition of Basics (Strauss and Corbin 1998) reads as less prescriptive and aims to lead readers to a new way of thinking about their research and about the world.” (Charmaz 2000: 512)

    +

    Corbin and Strauss (1990) “gained readers but lost the sense of emergence and open-ended character of Strauss’s earlier volume and much of his empirical work. The improved and more accessible second edition of Basics (Strauss and Corbin 1998) reads as less prescriptive and aims to lead readers to a new way of thinking about their research and about the world.” (Kathy Charmaz 2000: 512)

    Strauss apparently became more insistent that grounded theory should be more verificational in nature in personal communications.

    Glaser (1992) responded to Strauss and Corbin (1990), repudiating what he perceived as forcing preconceived questions and frameworks on the data. Glaser considered it better to allow theory to “emerge” from the data, i.e. to let the data speak for themselves.

    -

    Charmaz identifies these two approaches as having a lot in common: hey both advocate for mitigating factors that would hinder objectivity and minimize intrusion of the researcher’s subjectivity, and they are both embedded in positivist attitudes, with a researcher sitting outside the observed reality; Glaser exemplifies these through discovering and coding data, and using systematic comparative methods, whereas Strauss and Corbin maintain a similar distance through their analytical questions, hypotheses and methodological applications. They both engage in “silent authorship” and usually write about their data as distant experts (Charmaz and Mitchell 1996).

    +

    Charmaz identifies these two approaches as having a lot in common: hey both advocate for mitigating factors that would hinder objectivity and minimize intrusion of the researcher’s subjectivity, and they are both embedded in positivist attitudes, with a researcher sitting outside the observed reality; Glaser exemplifies these through discovering and coding data, and using systematic comparative methods, whereas Strauss and Corbin maintain a similar distance through their analytical questions, hypotheses and methodological applications. They both engage in “silent authorship” and usually write about their data as distant experts (Kathy Charmaz and Mitchell 1996).

    Constuctivist Grounded Theory

    @@ -605,7 +613,7 @@

    Constuctivis

    Line-by-line coding sharpens our use of sensitizing concepts — that is, those background ideas that inform the overall research problem. Sensitizing concepts offer eays of seeing, organizing, and understanding experience; they are embedded in our disciplinary emphases and perspectival proclivities. Although sensitizing conceots may deepen perception, they provide starting points for building analysis, not ending points for evading it. We may use sensitizing concepts only as points of departure from which to study the data.

    -

    Much of the rest of the Charmaz (2000) paper is an overview of coding and memoing methods, as well as theoretical sampling. The emphasis is on situating these techniques in the Glaser / Strauss and Corbin debate, and it will be better to refer to Charmaz (2014) for in-depth notes on these techniques.

    +

    Much of the rest of the Kathy Charmaz (2000) paper is an overview of coding and memoing methods, as well as theoretical sampling. The emphasis is on situating these techniques in the Glaser / Strauss and Corbin debate, and it will be better to refer to Kathy Charmaz (2014) for in-depth notes on these techniques.

    Charmaz (2000: 521-522) provides an apt account of a significant critique of grounded theory, and poses her constructivist approach as a potential means of resolving it. Specifically, she refers to the notion that grounded theory (as traditionally conceived by both Glaser and Strauss and Corbin) “fractures” the data, making them easier to digest in an analytical sense, but also making it more difficult to engage with in a holistic manner. This is precisely the point of the original approach, to present qualitative data as data — as conceived and valued by quantitative researchers, i.e. as discrete, corpuscular, disembodied, re-arrangable and distant entities. The text of these two large paragraphs is copied here:

    Conrad (1990) and Riessman (1990) suggest that “fracturing the data” in grounded theory research might limit understanding because grounded theorists aim for analysis rather than the portrayal of subjects’ experience in its fullness. From a grounded theory perspective, fracturing the data means creating codes and categories as the researcher defines themes within the data. Glaser and Strauss (1967) propose this strategy for several reasons: (a) to help the researcher avoid remaining immersed in anecdotes and stories, and subsequently unconsciously adopting subjects’ perspectives; (b) to prevent the researcher’s becoming immobilized and overwhelmed by voluminous data; and (c) to create a way for the researcher to organize and interpret data. However, criticisms of fracturing the data imply that grounded theory methods lead to separating the experience from the experiencing subject, the meaning from the story, and the viewer from the viewed. In short, the criticisms assume that the grounded theory method (a) limits entry into subjects’ worlds, and thus reduces understanding of their experience; (b) curtails representation of both the social world and subjective experience; (c) relies upon the viewer’s authority as expert observer; and (d) posits a set of objectivist procedures on which the analysis rests.

    @@ -666,11 +674,11 @@

    Rebuttals by Glaser

    -

    Glaser (2002) constitutes a rebuttal to Charmaz (2000). As Bryant (2003) points out in his response to Glaser (2002), it is very angry, polemical and irrational. I don’t want to go too in depth with the fundamental problems with Glaser’s response (see Bryant’s paper for the details), but the gist is that Glaser never really got the message about data being inherently constructed by researchers decisions, actions and circumstances. Glaser seems to continue believing in the inherent neutrality of data as a matter of faith.

    +

    Glaser (2002) constitutes a rebuttal to Kathy Charmaz (2000). As Bryant (2003) points out in his response to Glaser (2002), it is very angry, polemical and irrational. I don’t want to go too in depth with the fundamental problems with Glaser’s response (see Bryant’s paper for the details), but the gist is that Glaser never really got the message about data being inherently constructed by researchers decisions, actions and circumstances. Glaser seems to continue believing in the inherent neutrality of data as a matter of faith.

    This being said, Glaser (2002) did highlight the large emphasis on descriptive rather than explanatory potential in Charmaz’s approach. This aligns with my own apprehensions when I try to address the relevance of my work. I tend to use the term “articulate” as a way to frame my work as descriptive, but in a way that lends value, and this very fuzzy distinction between the power of identying the shapes and relationships among things and explaining their causes and effects in a generalizable way (i.e., theories, or explanations), still somehow troubles me. I wonder if Glaser is drawing a false distinction here, and through that, a false prioritization of explanation over description as a desired outcome. This would put my mind at ease, as would dismissing Glaser’s dismissal of people who simply don’t know how to do the “real” grounded theory (which, in his mind, include all feminist and critical researchers).

    -
    -

    On the utility of grounded theory

    +
    +

    On the functional and pragmatic roots of grounded theory

    I completely agree with this statement from Clarke (2003: 555):

    To address the needs and desires for empirical understandings of the complex and heterogeneous worlds emerging through new world orderings, new methods are requisite (Haraway 1999). I believe some such methods should be epistemologically/ ontologically based in the pragmatist soil that has historically nurtured symbolic interactionism and grounded theory. Through Mead, an interactionist grounded theory has always had the capacity to be distinctly perspectival in ways fully com patible with what are now understood as situated knowledges. This fundamental and always already postmodern edge of a grounded theory founded in symbolic interactionism makes it worth renovating.

    @@ -685,7 +693,7 @@

    On the <

    Theoretical sampling

    -

    See Charmaz (2000): 519-520.

    +

    See Kathy Charmaz (2000): 519-520.

    From Clarke (2003: 557):

    Unique to this approach has been, first, its requiring that analysis begin as soon as there are data. Coding begins immediately, and theorizing based on that coding does as well, however provisionally (Glaser 1978). Second, “sampling” is driven not necessarily (or not only) by attempts to be “representative” of some social body or population (or its heterogeneities) but especially and explicitly by theoretical con cerns that have emerged in the provisional analysis. Such “theoretical sampling” focuses on finding new data sources (persons or things) that can best explicitly ad dress specific theoretically interesting facets of the emergent analysis. Theoretical sampling has been integral to grounded theory from the outset, remains a fundamen tal strength of this analytic approach, and is crucial for the new situational analyses.

    @@ -696,12 +704,24 @@

    Theoretical sampling<

    Data Collection

    Interviews

    -

    See (Yin 2014: 110-113) See Becker (1998)

    +

    See

    From Charmaz (2000: 525):

    A constructivist approach necessitates a relationship with respondents in which they can cast their stories in their terms. It means listening to their stories with openness to feeling and experience. … Furthermore, one-shot interviewing lends itself to a partial, sanitized view of experience, cleaned up for public discourse. The very structure of an interview may preclude private thoughts and feelings from emerging. Such a structure reinforces whatever proclivities a respondent has to tell only the public version of the story. Researchers’ sustained involvement with research participants lessens these problems.

    Fontana and Frey (2000) spend some time writing about the emergence of an “interview society”, whereby interviews are commonly used to seek various forms of biographical information. They cite Gubrium and Holstein (1998), who noted that “the interview has become a means of contemporary storytelling, where persons divulge life accounts in response to interview inquiries”. They then go over a brief history of interviewing in the context of sociological research, which largely tracks the values underlying positivist and postmodernist transitions as you might expect.

    +

    Yin (2014: 110-113) differentiates between three kinds of interviews:

    +
      +
    • Prolonged interviews: Usually over two hours long, or over an extended period of time covering multiple sittings.

      +
      +

      You can ask interviewees about their interpretations and opinions about people and events or their insights, explanations, and meanings related to certain occurrences. You can then use such propositions as the basis for further inquiry, and the interviewee can suggest other persons for you to interview, as well as other sources of evidence. The more that an interviewee assists in this manner, the more that the role may be considered one of an “informant” rather than a participant. Key informants are often critical to the success of a case study. Such persons can provide you with insights into a matter and also give you access to other interviewees who may have corroboratory or contrary evidence.

      +
    • +
    • Shorter interviews: More focused, around one hour, open-ended and conversational but largely following the protocol.

      +
      +

      A major purpose of such an interview might simply be to corroborate certain findings that you already think have been established, but not to ask about other topics of a broader, open-ended nature. In this situation, the specific questions must be carefully worded, so that you appear genuinely naive about the topic and allow the interviewee to provide a fresh commentary about it; in contrast, if you ask leading questions, the corroboratory purpose of the interview will not have been served. … As an entirely different kind of example, your case study protocol might have called for you to pay specific attention to an interviewee’s personal rendition of an event. In this case, the interviewee’s perceptions and own sense of meaning are the material to be understood. … In both [situations], you need to minimize a methodological threat created by the conversational nature of the interview.

      +
    • +
    • Survey interviews: A more structured questionnaire. Usually works best as one component of multiple sources of evidence.

    • +

    Structured interviewing

    From Fontana and Frey (2000: 649-651):

    @@ -903,12 +923,12 @@

    Coding

    These notes are largely derived from my reading of Saldaña (2016), provides a practical overview of what coding entails and specific methods and techniques.

    Coding as component of knowledge construction:

      -
    • Coding is an intermediate step, “the”critical link” between data collection and their explanation or meaning” (Charmaz (2001), as quoted in Saldaña (2016): 4)
    • -
    • “coding is usually a mixture of data [summation] and data complication … breaking the data apart in analytically relevant ways in order to ead toward further questions about the data” (Coffey and Atkinson (1996): 29-31, as quoted and edited by Saldaña (2016): 9) +
    • Coding is an intermediate step, “the”critical link” between data collection and their explanation or meaning” (from Kathy Charmaz 2001; as quoted in Saldaña 2016: 4).
    • +
    • “coding is usually a mixture of data [summation] and data complication … breaking the data apart in analytically relevant ways in order to ead toward further questions about the data” (from Coffey and Atkinson 1996: 29-31; as quoted and edited in Saldaña 2016: 9).
      • This relates to the paired notions of decodng when we reflect on a passage to decipher its core meaning, and encoding when we determine its appropriate code and label it (Saldaña 2016: 5).
    • -
    • Coding “generates the bones of your analysis. … [I]ntegration will assemble those bones into a working skeleton” (Charmaz (2014): 113, quoted in Saldaña (2016): 9)
    • +
    • Coding “generates the bones of your analysis. … [I]ntegration will assemble those bones into a working skeleton” (from Kathy Charmaz 2014: 113; as quoted in Saldaña 2016: 9).
    • To codify is to arrange things in a systematic order, to make something part of a system or classification, to categorize
      • What I sometimes refer to as arranging the code tree
      • @@ -928,8 +948,8 @@

        Pre-coding technique
      • Semantic markup
        • Bold, italics, underline, highlight
        • -
        • Meant to identify “codable moments” worthy of attention (Boyatzis (1998), as referenced in Saldaña (2016): 20)
        • -
        • Relates to Saldaña (2016): 22’s prompt: “what strikes you?”
        • +
        • Meant to identify “codable moments” worthy of attention (Boyatzis 1998; as referenced in Saldaña 2016: 20)
        • +
        • Relates to Saldaña’s (2016: 22) prompt: “what strikes you?”
      • Preliminary jottings
          @@ -953,19 +973,9 @@

          Pre-coding technique
        • How is what is going on here similar to, or different from, other incidents or events recorded elsewhere in the fieldnotes?
        • What is the broader import or significance of this incident or event? What is it a case of?
        -
        -
        -
        - -
        -
        -Note -
        -
        -
        -

        Create a qc issue for sub-documents, for identifying sections of a document that are especially relevant and hiding less relevant sections. I don’t necessarily want to delete these lines, but I may want to hide them from view. Maybe this is possible using vscode, outside of qc (see https://stackoverflow.com/a/72954133).

        -
        -
        +
    +
    +

    Coding frameworks

    From Clarke (2003: 558) on process coding:

    In a traditional grounded theory study, the key or basic social process is typically articulated in gerund form connoting ongoing action at an abstract level. Around this basic process are then constellated the particular and distinctive conditions, strategies, actions, and practices engaged in by human and nonhuman actors in volved with/in the process and their consequences. For example, subprocesses of disciplining the scientific study of reproduction include formalizing a scientific disci pline, gleaning fiscal support for research, producing contraceptives and other techno scientific products, and handling any social controversies the science provokes (such as cloning and stem cell research).

    @@ -1036,6 +1046,72 @@

    Preliminary analyses<
  • Putting information in chronological order or using some other temporal scheme
  • Yin (2014: 135) also emphasizes memo-writing as a core strategy at this stage, citing Corbin and Strauss (2014). These memos should include hints, clues and suggestions that simply put into writing any preliminary interpretation, essentially conceptualizing your data. He uses the specific example of shower thoughts.

    +
    +

    Saldaña (2016) dedicates chapter 6 to “post-coding and pre-writing transitions”. He frames these as a series of strategies to help crystallize the analytical work and springboard into written documents and reports.

    +
    +

    Top-10 lists

    +

    Saldaña (2016: 274-275) suggests coming up with a list of the ten top quotes or passages as a potentially useful “focusing strategy”. Identify the passages (no longer than half a page each) that strike me as the most vivid and/or representational of my study. He suggests reflecting on the content of these items and arranging them in various orders to discover different ways of structuring or outlining the write-up of the research story. He provides some examples of orders to consider:

    +
      +
    • chronologically
    • +
    • hierarchically
    • +
    • telescopically
    • +
    • episodically
    • +
    • narratively
    • +
    • from the expository to the dramatic
    • +
    • from the mundane to the insightful
    • +
    • from the smallest detail to the bigger picture
    • +
    +
    +
    +

    The study’s “trinity”

    +

    Saldaña (2016: 275-276) suggests identifying the three (and only three) major codes, categories, themes and/or concepts generated thus far that strike me or stand out in my study. He suggests identifying which one is dominant, how does this status relate to or impact the other codes or concepts, and generally trace the relationships between these ideas. He suggests plotting them as a venn diagram to identify what aspects overlap across items, and to label those areas of overlap — although he does not mention this explicitly, I imagine these points of overlap represent the synthesis of new emergent ideas.

    +
    +
    +

    Codeweaving

    +

    Saldaña (2016: 276) addresses codeweaving as a viable strategy, but I don’t actually think of it as that useful for my purposes. Seems a but contrived and has lots of potential to be over-engineered.

    +
    +
    +

    From coding to theorizing

    +

    For Saldaña (2016: 278), the stage at which he finds theories emerging in his mind is when he starts coming up with categories of categories. At this point, a level of abstraction occurs that transcends the particulars of a study, enabling generalizable transfer to other comparable contexts.

    +

    Saldaña (2016: 278-279) identifies a few structures through which these categories of categories might emerge:

    +
      +
    • Superordindate and Subordinate Arrangements: Arrange categories as an outline, which suggests discrete linearity and classification. Supercategories and subcategories are “ranked” with numbers or capital letters.
    • +
    • Taxonomy: Categories and their subcategories are grouped but without any inferred hierarchy; each category seems to have equal weight.
    • +
    • Hierarchy: Categories are ordered from most to least in some manner, i.e. frequency, importance, impact, etc.
    • +
    • Overlap: Some categories share particular features with others while retaining their unique properties.
    • +
    • Sequential Order: The action suggested by categories progresses in a linear manner.
    • +
    • Concurrency: Two or more categories operate simultaneously to influence and affect a third.
    • +
    • Domino Effects: Categories cascade forward in multiple pathways.
    • +
    • Networks: Categories interact and interplay in complex pathways to suggest interrelationship.
    • +
    +

    The arrows connecting categories are meaningful in their own right. Saldaña (2016: 280) references Urquhart (2013) who states that category relationships are necessary to develop assertions, propositions, hypotheses and theories. He suggests inserting words or phrases between categories that plausibly establishes their connections, as suggested by the data and analytical memos. Saldaña (2016: 280-281) lists several possible connectors:

    +
      +
    • accelerates
    • +
    • contributes toward
    • +
    • depends on the types of
    • +
    • drives
    • +
    • facilitates
    • +
    • harnesses
    • +
    • increases
    • +
    • increases the difficulty of
    • +
    • is affected by
    • +
    • is essential for
    • +
    • is necessary for
    • +
    • provides
    • +
    • reconciles
    • +
    • reduces
    • +
    • results in
    • +
    • results in achieving
    • +
    • triggers
    • +
    • varies according to
    • +
    • will help to
    • +
    +

    Moreover, Saldaña (2016: 281) suggests that if you end up with categories as nouns or noun phrases, it could be helpful to transform them into gerund phrases. This will help get a better sense of process and action between catagories.

    +
    +
    +

    Findings “at a glance”

    +

    Following Henwood and Pidgeon (2003), Saldaña (2016: 283) suggests creating a tri-column chart that outlines the findings and the sources of evidence and reasoning that underlie them. See the specific page for a good example.

    +

    The constant comparative method

    @@ -1125,11 +1201,38 @@

    Writing

  • Theoretical notes: Hunches, hypotheses, poststructuralist connections, critiques of what I’m doing/thinking/seeing, keeping me from getting hooked on one view of reality.
  • Personal notes: Uncensored feeling statements about the research, the people I’m talking to, my doubts, anxieties, pleasures. These can also be great sources of hypotheses (for example, if I am feeling anxious, others I’m speaking with might feel the same way, which provides a string to tug on).
  • -

    See Mitchell and Charmaz (1996)

    +

    From Charmaz and Mitchell (1996: 286):

    +
    +

    For the most part, social science researchers are not expected to speak, and if they do, we need not listen. While positivism and postmodernism claim to offer open forums, both are suspicious of authors’ voices outside of prescribed forms. Extremists in both camps find corruption in speech. It lacks objectivity and value neutrality in the positivist idiom; it expresses racist, Eurocentric, and phallocentric oppression in the postmodern view. Our point is this: There is merit in humility and deference to subjects’ views, and there is merit in systematic and reasoned discourse. But there is also merit in audible authorship.

    +

    See Charmaz (2000: 526-528) See Charmaz (2014: 183-188) on “pre-writing” techniques, including clustering and freewriting, especially in context of memo-writing.

    +
    +

    Writing techniques

    +
    +

    Analytical storylining

    +

    Following Charmaz (2010: 201), Saldaña (2016: 287) suggests using processual words that suggest the unfolding of events through time that “reproduce the tempo and mood of the [researched] experience”. However, he cautions that this is not suitable for every situation, and to be aware of how this framing differs from positivist notions of cause and effect.

    +
    +
    +

    One thing at a time

    +

    Saldaña (2016: 287) suggests documenting each category, theme or concept one at a time as a wat of disentangling very interrelated ideas. This helps maintain focused as a writer, and may help maintain a more focused reading experience too.

    +
    +
    +

    Writing about methods

    +

    Saldaña (2016: 284-286) addresses some common tendencies and challenges with regards to writing about qualitative coding practices:

    +
    +

    Researchers provide brief storied accounts of what happened “backstage” at computer terminals and during intensive work sessions. After a description of the participants and the particular data collected for a study, descriptions of coding and analysis procedures generally include: references to the literature that guided the analytic work; qualitative data organization and management strategies for the project; the particular coding and data analytic methods employed and their general outcomes; and the types of CAQDAS programs and functions used. Some authors may include accompanying tables or figures that illustrate the codes or major categories constructed. Collaborative projects usually explain team coding processes and procedures for reaching intercoder agreement or con-sensus. Some authors may also provide brief confessional anecdotes that highlight any analytic dilemmas they may have encountered.

    +
    +

    Saldaña (2016: 286) then suggests emphasizing for readers through introductory phrases and italicized text the major outcomes of the analysis to help provide guidance for readers to grasp the headlines of your research story. For example:

    +
      +
    • “The three major categories constructed from transcript analysis are …”
    • +
    • “The core category that led to the groundedtheory is ..,” “The key assertion of this study is …”
    • +
    • “A theory I propose is …”
    • +
    +
    +
    @@ -1140,9 +1243,6 @@

    Writing

    Baszanger, Isabelle, and Nicolas Dodier. 1997. “Ethnography: Relating the Part to the Whole.” Qualitative Research: Theory, Method and Practice 2:9–34.
    -
    -Becker, Howard Saul. 1998. Tricks of the Trade: How to Think about Your Research While You’re Doing It. http://site.ebrary.com/id/10991105. -
    Birks, Melanie, and Jane Mills. 2015. Grounded Theory: A Practical Guide. SAGE. https://books.google.com?id=YsGICwAAQBAJ.
    @@ -1164,6 +1264,9 @@

    Writing

    Bury, M. R. 1986. “Social Constructionism and the Development of Medical Sociology.” Sociology of Health & Illness 8 (2): 137–69. https://doi.org/10.1111/1467-9566.ep11340129.
    +
    +Charmaz, K. 2010. “Grounded Theory: Objectivist and Constructivist Methods. Dans W. Luttrell (Éd.), Qualitative Educational Research: Readings in Reflexive Methodology and Transformative Practice (Pp. 183-207).” New York, NY: Routledge. +
    Charmaz, Kathy. 1995. “Grounded Theory.” In Rethinking Methods in Psychology, edited by Jonathan A. Smith, Rom Harré, and Luk van Langenhove, 27–49. London: Sage Publications.
    @@ -1224,6 +1327,9 @@

    Writing

    Gubrium, Jaber F., and James A. Holstein. 1998. “Narrative Practice and the Coherence of Personal Stories.” The Sociological Quarterly 39 (1): 163–87. https://doi.org/10.1111/j.1533-8525.1998.tb02354.x.
    +
    +Henwood, Karen, and Nick Pidgeon. 2003. “Grounded Theory in Psychological Research.” https://psycnet.apa.org/record/2003-04376-008. +
    Huberman, A. Michael, and Matthew B. Miles. 1994. “Data Management and Analysis Methods.” In Handbook of Qualitative Research, edited by Norman K. Denzin and Yvonna S. Lincoln, 1st ed., 428–44. Thousand Oaks, CA: SAGE. https://psycnet.apa.org/record/1994-98625-026.
    @@ -1266,9 +1372,6 @@

    Writing

    Mishler, Elliot G. 1981. “The Social Construction of Illness.” Social Contexts of Health, Illness, and Patient Care, 141–68.
    -
    -Mitchell, Richard G., and Kathy Charmaz. 1996. “Telling Tales, Writing Stories: Postmodernist Visions and Realist Images in Ethnographic Writing.” Journal of Contemporary Ethnography 25 (1): 144–66. https://doi.org/10.1177/089124196025001008. -
    Montgomery, P., and P. H. Bailey. 2007. “Field Notes and Theoretical Memos in Grounded Theory.” Western Journal of Nursing Research 29 (1): 65–79. https://doi.org/10.1177/0193945906292557.
    @@ -1288,7 +1391,7 @@

    Writing

    Patton, Michael Quinn. 2014. “What to Observe: Sensitizing Concepts.” In Qualitative Research and Evaluation Methods, 4th ed., 358–63. SAGE Publications. https://www.academia.edu/download/38719540/Sensitizing_concepts.pdf.
    -Ragin, C C. 1999. “The Distinctiveness of Case-Oriented Research.” Health Services Research 34 (December):1137–51. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC1089057/. +Ragin, Charles C. 1999. “The Distinctiveness of Case-Oriented Research.” Health Services Research 34 (December):1137–51. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC1089057/.
    Ragin, Charles C., and Howard Saul Becker, eds. 1992. What Is a Case? Exploring the Foundations of Social Inquiry. Cambridge [England] ; New York, NY, USA: Cambridge University Press. @@ -1353,6 +1456,9 @@

    Writing

    Turner, BarryA. 1981. “Some Practical Aspects of Qualitative Data Analysis: One Way of Organising the Cognitive Processes Associated with the Generation of Grounded Theory.” Quality and Quantity 15 (3). https://doi.org/10.1007/BF00164639.
    +
    +Urquhart, Cathy. 2013. Grounded Theory for Qualitative Research: A Practical Guide. 1 Oliver’s Yard, 55 City Road London EC1Y 1SP: SAGE Publications, Ltd. https://doi.org/10.4135/9781526402196. +
    Van den Hoonaard, Will C. 1997. Working with Sensitizing Concepts: Analytical Field Research. Qualitative Research Methods, v. 41. Thousand Oaks: Sage Publications.
    diff --git a/notes/potential-cases.html b/notes/potential-cases.html index fb2d59a..14e0938 100644 --- a/notes/potential-cases.html +++ b/notes/potential-cases.html @@ -248,7 +248,7 @@

    Potential cases

    Modified
    -

    February 4, 2025

    +

    February 21, 2025

    diff --git a/posts.html b/posts.html index bcbf026..dace74c 100644 --- a/posts.html +++ b/posts.html @@ -296,7 +296,18 @@

    Blog

    - + + +Feb 21, 2025 + + +Week notes (2025-W08) + + +week notes + + + Feb 14, 2025 @@ -307,7 +318,7 @@

    Blog

    week notes - + Feb 7, 2025 @@ -318,7 +329,7 @@

    Blog

    week notes - + Jan 31, 2025 @@ -329,7 +340,7 @@

    Blog

    week notes - + Jan 25, 2025 @@ -340,7 +351,7 @@

    Blog

    week notes - + Jan 24, 2025 @@ -351,7 +362,7 @@

    Blog

    AI / LLM, Methods, QDA - + Jan 18, 2025 @@ -362,7 +373,7 @@

    Blog

    week notes - + Dec 18, 2024 @@ -373,7 +384,7 @@

    Blog

    website - + Dec 9, 2024 @@ -384,7 +395,7 @@

    Blog

    meeting notes, general thoughts - + Dec 9, 2024 diff --git a/posts.xml b/posts.xml index 4ef3796..059fb86 100644 --- a/posts.xml +++ b/posts.xml @@ -10,7 +10,31 @@ quarto-1.6.39 -Fri, 14 Feb 2025 05:00:00 GMT +Fri, 21 Feb 2025 05:00:00 GMT + + Week notes (2025-W08) + https://zackbatist.info/CITF-Postdoc/posts/weeknotes-2025-W08.html + Some highlights from this week:

    +
      +
    • I met with Isabel for our biweekly update. We talked about challenges involved in introducing my work to epidemiologists, and practiced some techniques for communicating what my project is about.
    • +
    • I made some revisions to my ethics protocol in response to the feedback provided by the IRB. Currently trying to resolve a technical problem with the online portal that processes these applications.
    • +
    • I submitted a peer-review.
    • +
    • I followed up on some translation work for SNAP.
    • +
    • I continued with my methodology notes.
    • +
    + + + + ]]>
    + week notes + https://zackbatist.info/CITF-Postdoc/posts/weeknotes-2025-W08.html + Fri, 21 Feb 2025 05:00:00 GMT +
    Week notes (2025-W07) https://zackbatist.info/CITF-Postdoc/posts/weeknotes-2025-W07.html diff --git a/posts/weeknotes-2025-W08.html b/posts/weeknotes-2025-W08.html new file mode 100644 index 0000000..9d9388d --- /dev/null +++ b/posts/weeknotes-2025-W08.html @@ -0,0 +1,825 @@ + + + + + + + + + + +Week notes (2025-W08) – CITF-Postdoc + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
    +
    + +
    + +
    + + +
    + + + +
    + +
    +
    +

    Week notes (2025-W08)

    +
    +
    week notes
    +
    +
    + + + +
    + + +
    +
    Published
    +
    +

    February 21, 2025

    +
    +
    + + +
    + + + +
    + + +

    Some highlights from this week:

    +
      +
    • I met with Isabel for our biweekly update. We talked about challenges involved in introducing my work to epidemiologists, and practiced some techniques for communicating what my project is about.
    • +
    • I made some revisions to my ethics protocol in response to the feedback provided by the IRB. Currently trying to resolve a technical problem with the online portal that processes these applications.
    • +
    • I submitted a peer-review.
    • +
    • I followed up on some translation work for SNAP.
    • +
    • I continued with my methodology notes.
    • +
    + + + +
    + +
    + + + + + + \ No newline at end of file diff --git a/research-protocol.pdf b/research-protocol.pdf index 3821d1385919df6c25dac428cc88c955e395c39b..e18b2e47fe50f05370f920773f5399244430c58e 100644 GIT binary patch delta 8722 zcmaiYMN}LN6J&6AcLIaEYjAh>!DVm>ZWA1WI|O%k4G>&|yE_DDaF_kwf7(6my)Ru= zue!P}N?=b)VC%bs!Q3Ju$lu-Ftjrvdz1M(An|2{$m|>8BpMR7X`oMf}ScSBhpf&7d z1a`qv2_q{W0-x8@+^+;cz1{Up0ih8xA)y0=MV0=K*#PYsz%QM}U&9{-OxU^|Eq~p< zskbc-T6SyeqNnAMjqg^SgKDSdHcxIoe3%t@I;I3KAonkB;H}QY@(S^P&jJZo85lm- zq;#j~l_%1qh7v1f`rg@hn$@iy;F7k(bHn9gLl2xPMV*Y~Jq}c_Q#s}ShNQvK-=LaP zd*X;jDqtmvegjvX)ZnWR4SvOo$;MHmz2qc!9M*#s6BuRS**ofSv*iE)Y!+p4<*70;Jtow`Z zu+5O%D|2*NOb4Up2#dqQKHNDoS!dS1F(Oc}a^eF97VdmNHMt=QH(4i?4$S}mkddfv zAHt3~vi*+hb51C2VbB?bSW~jI>^h^>P`+H z8Ru%ASMC5?c?evEF3zakA36mMVX!_N1`rl7Da|)=qY+u1{nisx$E3W_c3uHb%|z415FjRh)|HE*A#Ry9@ah1 zLO_$afpb*3h7|wO{(GPpWMZC>M}QBEvlapW)H4U4i@v_)i+3qnb&y>3KO|?vXrVbc zlApt0{Gjci8=~J!e<@kL~kuyPx7AW z7hXU2G{+m-B%j-k!!N8?o%593xZ<0C9^`Y=q)8BQ%m(YhyGG-L`f&QxGbcUpeSAGL5QZZ=4GlbkAkuN2MO`IaZ?TZSz$9Mn zNPEjwL=Q<;qL!9@0jW6csylaInJY(?LjD}XN?H|Cok_vUa+`*DEogl`Hc5{Lj#{@eJvWXlNH~# zjpfY1H+BpJV~=eveBM(R)fj;exKiHALS`@K15`B&`1mo zPS~iybrNBhEZiHMKp?>^yt(M~dH3-b&AIQ@>iA|M3e4Q(P3M#sH*Je#;|meft&O{P z1iSNVRn8|^!BYNmbaQ`z&-ydNBJ#9Vh6gG+HJiuvI9^ER&?Qw9h?LKba2;zTTZWb$ z;5R7s1?UUwqU#Ijs;EVQA*#_;7f)67I}mrW!aU?)8l8G=OO*u|6$Ckp8DBmNH!MR4>wJ7zfK15eP2LGg{jk&>`+ z3?5b4kQXUsi_}ma?F<9=C;@C@I9(sXvvM`c<$JUU!p|}{be}B8?*W>SotLvJ>k9du zL24UM9qN<`q#7!->~L?kD_J5;8&=Z?rW;%At{!9_*&O5snK9OSVq*F4QB#%;NtQFZ%V3%l~+q2V|K1ljpuwaLTgpen`(M-4%=eS+~n@; z87_IaZHnKT_nca`7)LWLXVB=p!~u;x=*-pte0L{fjC?T~l0^8I@-B%$F1hThWN|_{ zi5xCK6*r%|IE@7t>vyOWkUs1xp|IAZfB;dpqI^8pTXmbV%l<2)qS~n1vJm2ZXw%eH zZWv|@VF4-z|KD=vv~Ly6GHgYh)P8Q*MDO$2p;(1!F(0;e91?|MImBTxamo>g)L}06 z(#LsNaK4t7s=k(kSmdb4@EGuL17jq(*j-sIcR5?Cx(1h*{=bh3-m=F|FE_ZU0^Q!P zCcn?QKYVYG2+U}Dd!;Bhc!2y**K^TyP+Z9`(G=h)WwC^28Wj<-w53|gv`}V>x5pMh zvLpc@V{wB0S8*wt_MG^Q1&{NrI9qBx|H-ObC^|xgTv|xNh9mukzhyY!J~fk#mkNMZ zlKyZV_-2bZ5~F zS-0);X4`*CCN_O6U$3{dvB>2wbW2F8ME2BI8fobrV~HJ`GcNwe`M!Kuok}dPG2Rpf ze4)4S|Mn(`r33T+KZiGY+-`s!Q~L1(nZ;1yn!0g8g1uNp12uqF`7usd9<9Lv*Cw&2 zu3E@lT@qTBdUC^O$f~85tNSI}RKm)9LyP*as8Swk%_lzKH!0U= z0}N*mSt?`y^y$v&&h%yF?L}VY%c=$uaLd2I#}<`T(D9FPljR_(VI!8d>8 zY3i?gv5x`do7x-9zXDa;=2tuIM_V4~ojebDvmMw5TngjP>66OogFA6gm9}Sm%F%wW zku&_F`(PUMzv!W&T7PR&46}I+d9K)W{}ynwoGeu;ozIy7ZAIij1cE1-hi)Z81RLYx zbWPyrTTy}uL1A9l&BH=7%wa4DUpve->KPKkW20I?nD%CS-WZf|2uVmOEaIf^Sc9{f z37t|h*aizYhz5Y0+o0gss8m;2o#9Z%F+4HWDXM8ixL}|zVCiN|i#tk?_@e;j(e%H?>ABp( zLZgcAHqzM&vGw5mWTg+j0M+{d*u?RSv<`SbFxo=K1l5NV;17RD!-qC3gRlum3p^QI zBuJC(0j1Ds2CA{$IW@JC6kHP@mHph79a;xj-KBUbg9ZFvs3fObfxHL*AWI2>XtLiy%fF1uZaG~Q^=!243oXyYn{IAWsgt{OU z8?|81_!Wjw>PRzOSgAc!BSHoj_}*L4u5=QsgYb_bIkEc8# zI*z|lJHMVipZVP1_3=Ue-QwvimuxtwMC%z0a(?Mqh?FomozsBShul5ZI3qVz%cyOE zEj>LvX0v+EfUMy~i04Zh$G8inhe0?K1!d@E~?-rXMu^oJAKSt=U^K777hSFVa zEf6tjl2Q7VDl&7a6I#VrzO9=I{?4dQha1%c>jwEbXigW;@$8cQ>9QuvuKHX=Y~W=e0}bY!pK)F~eiE zuyybH&VIsTVWk)i4f9+l6&1xOnl@E$P-fEEIX47qul5`FMNUh8u6{c-?@H|r;Ywni zGbKt$O;}M_a!C2&@dA`*¾ijaTD!m{Z875{IZj3H*O!~!9--CEsDFA4+&E}Mgv ztqB~Ln)B8p;T*WO^lW9G)+Lw^iHav8ls)@3L2v=i?}ZOq38&>^)G?iy{fc01-@&=h ztMX#uOui;DS5tViIRP=##i zyV&nwNKK5~p1hrFvH@4Iq@fNRf@kz$-U?>$sbIGPt|9|3Yd1Z9Y9HRXL9ta69D)MX z-@ga8WT2Jd`3mjwn!^DLSHG}IRjd&E%6!{`Z7k`0y?HLx;K0joiXiL&A6%r-QrD;G z2yV#+=A!B9MhiwY7n^9;xM93!@Ri@PDO*=eh_$v!t6oX~3N2V?2?C=tK5YGOZ&}J4RsaMCO{A zh30i-O?cS8GjoizMVL;utN1Y5K}(1@q_$Kvtr<{Xi0aHt>D4eWlK8dntC-K@bE_?I z=+nvL!CORLF6Ee~Io*DQf$gZN16KjDd3W8LlxL6Kt*Mn+MB@*bAVhgGCH0itO4Hy; zq2MZU$|)7HLpYNsE*RN0AB&0y8(755;|1w+)Ru=Zov585kC3!M2dSVmX{Xr8U5B9; zqT<&e;f@Fe65p4lU5e1)h-=O9UkIDmW6Qz!@sqUX)N^MH9U&n45r((rZ~f%OyBHq# zOJzB=yP57F2ETym-D6NZI6rzcZ*CksBy#e`9MOtRvoEo z%^w0WK*56u&|l2?kW7symw2~=I&SSdsE@LM1%>vl2kRt>=nCka z%hPk*jws_l?-TR;^SV0UTCM8)2l3q>=8z~I?k;~9IBIadeP=qdxf`!OTG^2<2!pmt zK1cKJ=`HmB0+Vn< zLHFje9r=ne6oBIVefTpjxI{1-je=3DPSVM7*G6u4^xtJSO~`a~Z$!3w7GZ4!b(-!H zJ8HovM?DyXYt{9Nd^BuMgmF-%?4{uN`>nlY@xNan1-Bd}oM0eZMN3TF7tPq_IAP#lu*M$Qa!8-*7-@n!N_C5VKJ^3 z3+&jl=&|4oko=ufx0vvVb_s@DGZ1T^b$L?oMD@wx{faX3X7|}6(*SggdAL62y&50I z4mNR=iwF(AGapC<<|;=h5+fpg_q%>~=+=ulU9{v4#XbzGX-&J%Je~^NT0WL;n99AX zc)Xk;H|zf5BGrC!CCOru!ZOQBJAcBd|K6dNb>d%O76BP@xwb{tj>72E;SUN+RQ z7Js1X5ZirlEJP7DobqBKHq@6Acve>46JN?JRH z_3X~*A44B3OWH)gf}1&k zQp%F^7qi@S{-TzWFjrF?f#0%JVT9yrABi93B<2kqY(sIne>FZM4IWEQbxwHsexOv^ z=d80_EWvi`O9rnic3D5xic_pz00%yrMN6N6ORJ%Fj*S`d27Dtf?i97s7ds;PKx$98 z&6iF{jM*0gu#tW&GyZ&)g(3EjuPNe*zUvB##b z505(b`-XlVE;Pt_wTnMn3b*P#{RYT)FFF>MhdI9(S}^aw zho5@{NSP3?3Y@Fyol#SH*=t-6MYu>=hFz9^UqRl|xFF^A|H6y`5-hCHJuy4fng4r% z;V0rHlz*~xdKq9l5b?Q;2gcW0BY1H+?D(?4x!|F#cxRa75Q{-$_cS#G#hInD@wb;m zVY64O1YI^^nRgNDM^w$mOS#}$HU_GR7k%^+KY9JWIFRC~Cce(e`78mnH=IO4=y_Q> z5UU9v(%Opij$Vr=RA8FI5dT!d+p(}I)>unYFlhRyU}1IA@l4E+2=4BP;NhP8Ho<2c ze0rc9x#pCrCvs|nU|k$vcJJ1rRHsz7CA7(vQC)hk3pw?Igt9-~qPKJpsq-8)JYP^; zcGop!xvY+25ZpVY8rPQKZWpDhLe9vo!+58o|HWUx)e3aGS6>sv)pH`~{n1;Xm{aU$ z>Fx-r8}s)UB;Vcd06&-WQq1?+Y?R0POEigQ`vrvB9RI6W7rwx{t?@YxKQf53@Gm*G zc36u6BJ4lNSXc*U7o1vy)=}Th!k4&Xer^6%e|_W_mfhK3ixCnI-8KM(v#`ncoXH#_ zXm8rV3jS{Md~s;htL60$6ftswt<0UMtLLuoA4hXe%<4VZ{4f6;zS(h9rb}P)-@f+} z9Lw?w{{f9o=sd-0aO%c9m+o@@LC+gHKDw+Q1@4kRZwzhx@Sm4*8fz=+^Sy6eOMIU zuw0wYWOsK2z>3k6ggl4qgw4j$E!_iClXgj`9l>2J z&$gY1P0HFo%$)&=e>z7E=^F{~Rtcidxxz#NK|K|)fne2O$|Y>eGh83>L>i_Wc2RuA z2xjq77o>kbDEl_m5=rfzZeUd**t(@G~yyuCn?4RKmz5X*Ff8kG_ zZvfZwWq+d{>uXDO4Bguy!R1&ts6-9S!avee`?u%HHGr(`LI%)j_LC;3j5U9XvVXFG zhFRf~G+3?F!VoTLd&V}7+VWS`;+JJUA@MmoanTm_exXCvFTm}oMOC;9eJ{cU3y(3r>^A0pQ;@H< z4b4^739*cX{M+yFZK?*h8z|Fpt2FTpHkDIAu9-GG6N}CR9whB}B8PgfO@nOh`tSKeG z=#@w;LZZBx^h^Modm*u-;G!{_`7uceyha89^)Bo-V@aW6h69O|!`cPd@n#s+eeFaE z4pA+vuYgIVhhMt`F4&T+p;s#sU$PCWe8ocvwimrJ@vb>lrX* zoVML4#n0B!N}*5M)Mc-427q9?JgsnWv85IF;;(S8Ai~F76Fu{n(EYpLnh)z zsB&Ovgn4}g9OiRBw1%C%fbTY)?$Q#aGH}4RX zzwPU8XG=E^kY=}tcTlW+;W=zs&?}9UNO+_*GML_h@klJzWS2mvdZW!^<0{5ep>Yt( zBZMMl6^KyMlPNBUdkThP|2xee8>!EibXnZ!)U4W3K)De6o%UpK=CJ}tvGyyV&%Rzh zz<;LO=kD|1jsg5#Oe;QOTrjrV`{rqyZTs!<7GjKMqsZVJe>Bfg*8S3Kv`A3ao`D9%0LD?Ht<0NSF6)d6^pd^@2~gEgl0r0!Z*yZxec>ts z54@H}VbZr3lACUYPi7;U@@V%L1=OsuYepuvzWpjGnf-2gglg%}Xn`qVurgWju@t z<>(In7R4lbWoS!w>5+ zZPL*_argm#&2JZ-N88}0RGAGrgEtX_J%R!n&%&WWUITYz<~IuAtIr0P5zxtk2?~t~=D0Z{=`)OyD z9z#VO2o94c1=ELOJ+`!s>1y9k@u8Ld&qF!ltU0xI#cKD8x>&rn=u{V!YWIX59-TXn z>h^Vu!;_PfTV3a+xzN|azs5dQ2L!2Y-JKn1m9cp4I%qn3n9rV81S+zkBF`_A zCWb^DXV>KQRQDn}-Q`^^n|d5(2uuYGdMm$;M-)_lX@DY5g;wBsjo%F0Pc*DZ_^b)| z+A7?IMz@Lib_Jc53NAmtlt^0sOUK59Vv!KA;IZamH?!j6=j7pI=i;~IWar~Dv*P69 z=H<8M)vA%8Jb}v6$YlQ7X;BO8GCg z1kHN9l+=rQk^@$WTAC%nWcznBrA&?w7=m{U){ZfOZ67C*+nl2=h!A8*Kvt@_XIpiU zI`#>X=~|$5kI6rXY$XnIg%&vYSWeFvfL4pm+&_o`t}~|#?1$o4qoJJCLGQOa?`PvD zeAn(zb~wcBs>MER8B^;Df+CC*=SK~3ZC5Z)ytpwX3*SLZU9k2O#CERuud)T;LiXrU z2WAQ5EL@o529Ws4=_(ICn&zqM$*EnjRfZY(@yg6n6bgDSVuum<$!vC%V*2z6qnkKX zN&wSfmBv?fqM8 z@3r|qhMHT7rR_p;5dXVQ}w>G7M9?hMtbfviumJb-MX@nDt-;oYWZk6DKFn&l49HEB8Cw z1^h*FhJD$Sb$123|D7c4wNGr&(X{!yU&fS#ntbUjA-8Brm^kyLsdowKiL{S;qJc1F zbwx{I4=a9#w)tpew<|r()NqPWF$8!6pO)irZ=A`zu@!xi+j@tqQc;40$0h}hlram_ zB)fr=OpRa!e2NImhvH^Hj*_Zkmab}OXxi@(oIyzrUTU*I1>l#m+kR)hi_Zv}F&u;g zT1J(W(2s?h$@s+O(`3wK4CY157$RL+b_xuKn$dNs=mmL?v0QUJd-wMV<%Qj-TO7%(Yu9yob5|Q`^BNmh30h>>QTSCKYT!ew5;$Q5i*d}8YTwAC7g60 zMgv3+y11NcVmr?|V=|Qe&nn;JUUTE4IT>_hvY6s_Z?W7P`_|3HMSkMX%S<=Qnz8yy zOw%^L)*7@i2O-`2(c-ZJmfFH6^IWhOpP4wp#q=&&;?35<<+RQ7`^er`cMf~rb|6aG zy7BBrub4}W3dYXp<8UX0Lh84zLRnr^b^NlW6sq62h^3Dc){OmbrZGWbnF3 zE~CF#wts2e9LM@e6eun{?$t42Mdj_B(ZReHYAj;Kgnye6y7p2%Yz&%>0rZGoP_2%!3zjb!G~9_nbsLZ0@+I!irp1j`{cC2tA=H*4(LuSA+ zYiRBRY}cJJE}M?LUN!k-MzRE%G8j>(+PE&<16HRanTN|<+n}*G?(_-c<$Ft>*eju5 z;f98bJC^;eYNFXuy`R26OpucR5z0Vrmk=(jFz>*jn2JiIA4VNi#i0=>6?XISY#Y%5 znks!Q(Zo-WLjvP7NscF?z1|YtZB=p2cm%tRT2<1hIq`qD%NYi+jLyTNwP8%SIum*gGoMOqy=)D!5LS2fylm9nFbjOve58SqkdCF8fYEkEZT zPbxtfg_J}g(SJUE;Fvg#1Xj47>Nk`pS)~!gHU#9G=F3v;o@__N-%U=NGPY5g@Wg#v zvHH_c7B|hfor;e`iR@64Gm!@`No)U;IH@*L1?aHHg$26$z!zy@qyI=IK?3pF{r(S3 zwSkAz73C#OV7^fJ+iUYZns`+o=u7UbXstry3f_95gF^Gu#mnipFP_dFYpV`G_Y_^m zpLQKoQMwyCMe(0{PQq~DNn_zR7GnrO-caD@(bL}1`p>P@jejQ_X3pj%KScu1X{zmi z0vVq6BnQnlui$<6+@#U}Q{~m7G)}qP8D9&MPILr9)Y$&>B5n@i36kMrsAge-LZs ze^GcS{-Qv()0hA4`kN+Kc9%4DU#FqU*2%j^$-(7R+E0q{tBfk=DK*+%#%0VcF-EoT zz2QHP!-Xe$KcNyT9!`?Fz@pInNl#!(U_}z1?EIJW_(ryBQ)xD2Ly4y=2PKAZ4n;0V zh6SINAvp^(o7|Q{Jqf#qj9W-iCYr2mJz*jp^N+sssJBUkN+q4#ijQJgnY%#8z}VKK zxLbL2ogQb-6%B9umgU3b^Iw*PqmxzGbNKqDdYcd6&*RM*>HqhD*TbCf}6>A%TF8 zTcZ2sEq?8Un%%nApi7M)e$4L7+VE_Jq^K-FV`NFejv;d?T6Jg3;~s11%v8K_zN?rj7KQ7AHKwm`ey{lxn6UFH&pSjkok zlgc(>bQ%#wfI)8<+khCil8#ckJQ3zVlcAN0>8t9@ZY^A55FhUD!LfO)qq~P{Uzttt zG#UBVQ9WMja5B>yS$ZS^hU10fD4V?F5F5_&E`=06mhzpiMZuYky5)70I=Q2ER%N=K z&Y8lZM~IT=mIf7eFUBg%DdaB}e)xevqK8Y{#^=~O2cU!Zx%X@m# zJ%!|Mq`=rpeU1<*hU5$#o2zO9iqh)M+e>#VDV=mweufe}u_y{b$EZ$`yv0Znz{FYef5|;%Dua0Gs0Jf@tI{LR4ohlw(nnM&abVYbPQa?(cl<3?zjlr>$%yb_#4o(!Nv@6mC z5Oy#HN+qI6bS0FoGK>NcjjRlXte)Se=&6GH14l`dP|Gbu(`Fw98TZRcJEyrEeI3d} zY}6=pM`5rM9Uq0R>Iw@Kr88oZXJIro`ZzQg4au^0mX1S6=0bZrRES|&Q5Bx8IhSNe z2k{*MoRyZuL5Rj!N+98BSB&{quP@dIB*S!4q!RRtkH>+8&|OMXOL^inS#GQ29m-wG z1%2CB$MaEO(U~olnZd9#w}Q}5X%jP4)_f<1OPu?9qO5IQT#&EK%-r)Y zdsy6|PwgMvVJumZ{+vKR631>jwWU((@a&*nauPXw1KEz}PMz+F;^2HOb2$#Ly-u%((h*!p!|9w@+2ix?ws z?yuGT;bEvBXVdsx>F0UxHDtb}YI6k{xgBaVyMs@RBhoR~En7?q)ZWbZ3$?DTyfV-I zch&c7@7exlto!~`$nVpuEt5_E!0N%-Ok6nNz3LIrV-j;zloXB$NxQdQjH^-A(sEGK!00}?c+LuLe*>vUdJtyS!4&b9FITk+h{NshSh5a_J>!Nqsd>T@Q z+2bdeS6kP?abV+aG&o0-3jI`~fH$X|mg*NyUfk;v@HO3kXNky*Ys{|0V!(z28BY=arMcvYuCq)ElZEQQqjuaag>s& zgadvZChUiV-SWWhs&MeEudsQxu5s&JLle?^x!U+B41Y36w=KG4_GYTDxHUS3ogD>j zzZZv|q4Lc@WWYWr@ZwULuxO{@lKqU?K>e`s-x2sm4+e8eu|qM$T1rm_z-rj21+#GC zP2XqLQb|b3aQ{Uv`q^G0Q^UKMOF1o;9BVgIKFor?5FVJVj7rx0B95!HXW=x8Hj=cu z+8}0bs*1@VVe^fw+JVQs2`CeU4pWb=W}{Q}JH1F2{<};#tBQbtz#@Y;LH@|VVx2d4wuL_7y<;kJ^-(W&iWhh*Ij4}CVP(@LU-f#(r?O$^thw>=hZT^ zjt1H&n9)_Zf<}ZV*#dSsq^{R(?DxN>oo9HE@q+^xGXu`?*|)-%Jkv?!)OZ4$;N0Y(=U#{usZm7FY~h0k1!@83Ulr(1;a668xs z^)Y02B4)-ZGNBbSJ5#Zpm+1ak`wHi9er0a#HmE)|bP;ZE%U`@6Xk~+CfBn0>b(-&p`W!etzwu9uXPR}e zBC|uAv%#{}$DJ=3 z;(_H`MoP?r^rm2Ff4Ld5UYWS5KUc7XA1mk#50gV+17Vtu-$f;tBW5$gWs3EF<2%)0 z-w04<2re&nUGp9MMjwwB>?+_WYGKC(p2oy_;6ihe?#6dwS&K<~WHDY?+&A~+|F{Ec zJyny>1|v#nON%1D_S2U0F`%2XxLoN{oj~-JiO6EKq&m6O9jhY@%AnkdIz2`)XOpQ* z5qV3lcPI{I%tTq}*@bI8~*%WWFzPq_4@5^I3^f>|C zpLgnu>$5ERz2^Urs(+MYB{7p|3&sY{G;gTDa|>rxf#z>nn-G&0i>H|V{6F09En)!^ zIvq4p+t1H+>*Y%H$`U?}HE2R4sO{)^Zq(9)TcQL?_1%Wp)JSN~C9J;QCh$J^GKpcW z!h4@mecCxgAWZzqW$M|B8&`l!brC&+sXNWPt z*0B~`mm~CP5CXS0bJ0Be{KJ*|O%SP7PPmR{&9pSjGXi_oGdjrTZiJ3xnJ@{Sdp7ZS z!#~I;7NSfUZ`(;TY+94$IhS|6`m!U9bue&y#?w4ALp{EnbkjhIYXR(8i^o8QKN$ki zc8X6+YvwJ zcfyT^nT!ifRZHis8(usM1m>S*1~9h3JG=9El{k$4h!)-vN8j+1N zQl%z#JjC|B_Nm1$*no*{)1n-VMUVVfBU1LB{X&hre;FoQ(PEKWP1r&D6ZFRsNr-7k zOb6696*USMlG=8Ug_6PGuLM4A9oPBO()2x?W_b{uDS0 zLfFo+eDf2?L-2|>x#GV}2rX@1nL}dq=eNGajhGG3_RhGvdqdP&7HKtJt48*31i`hY z`AlA#hDbE6cY@ySfT}m0RgK7tYgXKp&Canq#}X#swY~tUPK9$6q>Pxq-Eaw$T48+{ z6OdNphFvI5;L70n2^xqiX535vb_fD{zLvDWF`*iieVqCe1hUKjM#9+|leXT1k|i~xsym6d zkw4~qz4G`R7DM^@Sk~V_$;_7-AT)h$FGBuSa`g}nw8rYn*g!j*wCzBRoa$@9b5^+c zsB>qEx$qMjX?-}|R>qSsh2T5LcUxRMfW-UPJu7Nstf5+l4TfLQSI=G|diWaLR|@Ku zzQ|oJ74L#EcOphnHq?&l@ViRZgdLxt3@BziZzCZc5>v*M<{O(!)U$&z^pafXf7QZC%g_*R)UyL7=Vj%8srZ?M$+FZ5#k zQKZy3JFw3%%TVw3?)c7SsCLG=xp=jSY;k==ZXW>DkrTP~=@g@f;r8alCp7L1 zcv!UNuZo!GS|sofdGh1O(~o(nN%Ahz0-pgyOK!*gPmM*&Y8qyR>9Zv}M@&4x8N#hux zeQo32y^ zYzO=iL{B^Md}33O^bVK5M0p%N;y6Uu&th;>=^;m7f}~B>Ea>qHo0bAvnLP!?-FxkF z+|1h;4V)1%kxRYdC!DNQ>QYHtM_02F*ZC&i|Kbv~7qDiIi#^)NUyekTBCa0*Q+`&G zUXKUJsCfL8D|^XTcbC@Un=ISpBYOW^gWnFnO33AfeEa~1_7mnswxjtJwD0&DyM->D z1kMDk!#%hI^;T@Cu6Q3+}eCrj0>xK2|1^9eJ zep4o)yL`0Az$%aVjT-q)js6<=69=w7mCCqm6g7&?Fd_W8-ay0(Sq7xSg$Rdy2?~bm zWQ^$I_j8{#-1jl{@kXQ2=3{p+njg+uVlti(EbplvN!r zezb+1Pg(IkxeybsYndiwb~-m^MNluppSn-+gsuJ8cSp;Gl}+iBHykorTt{=Cd#E8i;8@QPv1~&@~OM$O$uHoFT~-lOM95#Cc^d))k<%&4^{fR1gjYDJc!6R6X*w z8N=%hYQ)`6y$W!Qg{T24^mvn0=RT zGX2QiH@-NL?awrtZ^o7rNhWu(ycPpa6Xp_0(~WQ9(54O!^=eoHBk536&OWySKgE1H z4Y{mD|G z)A3j4KO=j#uX5C*KXEy#zh4J+D9$kxQ6pjpCQYUh&!LpzwB%?+Lt&(4%QCRh&S`$} z?}H*b_@1pWSVgMjt!*f{#<-!0bvi7ZzmszR@+}pKRa;ks7I%KZTQ%i*ml_ACr-?rW> z-DjsIp+hqR5twinv+nLsoC~*1wZx5L6g>*q$??(a`^xKKG474KNr*R zyW2URwCWN;99J#8Phai)(s-xyvYSeYuHuBS$oy}_>dw|npc++6JPTubrvLiP8~QXa zKBiY{En_n6o)GY^fhGDW=IAxL?`aT179fzZL?Y@0%%nPQQh>)<42p4y_JBRRTT;FX z&p-+ujuh_t@1XJXdFSjXe*fSNAe2;z zsA{4X4*vODIqdAk10hG4SzWj=O-L}&Wi)>&^gV} z1`M;Q*_{iR07%aM@?iUXa&wZ!N$FCI{TNBm8<+Z&HG?BM3jmO^i@+0nU{XAN8njt9 zd7Re9q|PXSadru8PYQv?O@n8)+^t_Gw(`3DZ|1!n?Ebz7$sa6-wS3c(`6Ngq#DwMo z`nHvKvGe^46Vy8_EZq-+a2!#i-da{iWCISXds8S|3twgVz%o z-6COq5z+fw8^LPc^Km&L=wR0$0I;s@()}lx(%RG6fnFJd@1cjTw<|uFf^pwmN8@d8 zOz-k_-7CJAU2h?%ZPz6(MZe6YC-BZvftA)qd=pl$cynajg~i25P)~KGPuUiQz^i7t zelDpLZ>``3qa`SYuop$pw9;{J_oVjRj3C z!)3{DZN)FhZOy|YXk*RA&tuJNVadlYBxETp#4E%FkP!R76lwlg&|iQ;|L^)?rq(t=Q;@8}85dom8wL1>pn@gIWq(o|-1IK9Sd0ocB$88(c5j2w+?)1o?u9U+742nUy=m9#_K6IiuQf^|w9aobO2X5$T3+CJrR3B{4 z9cv)lFiuzS;|)ow1^g0#ercP%E9miBV6#slWe(wylymw=_0y%1;jz*EYV%>ZgVV2W zyBroziZh0kQk}Ch^+y9ih{vI^x9Iz1hx~y`)4fVrXO*#T&qvdx)(N=I`RlW;)BDtk uq_J`HR&RZ=YXT6fqU|-Tf*%ca+)`tU4g(QSRPcN^bfbu^^KkNVi diff --git a/search.json b/search.json index 7ec026e..7fbb80c 100644 --- a/search.json +++ b/search.json @@ -80,268 +80,240 @@ ] }, { - "objectID": "posts/weeknotes-2025-W07.html", - "href": "posts/weeknotes-2025-W07.html", - "title": "Week notes (2025-W07)", + "objectID": "posts/weeknotes-2025-W08.html", + "href": "posts/weeknotes-2025-W08.html", + "title": "Week notes (2025-W08)", "section": "", - "text": "This week I continued my methodological readings. Today I got some requests for minor modifications to my REB protocol, which will be rather painless to implement.\nI also attended the Public Health day, a student conference where grad students presented their practicum research. I really enjoyed engaging with the students at the poster session, I actually learned quite a lot.\nNext week I’ll try to round off my methodology readings in anticipation of my REB application being approved by the end of this month." + "text": "Some highlights from this week:\n\nI met with Isabel for our biweekly update. We talked about challenges involved in introducing my work to epidemiologists, and practiced some techniques for communicating what my project is about.\nI made some revisions to my ethics protocol in response to the feedback provided by the IRB. Currently trying to resolve a technical problem with the online portal that processes these applications.\nI submitted a peer-review.\nI followed up on some translation work for SNAP.\nI continued with my methodology notes." }, { - "objectID": "posts/weeknotes-2025-W05.html", - "href": "posts/weeknotes-2025-W05.html", - "title": "Week notes (2025-W05)", + "objectID": "posts/weeknotes-2025-W06.html", + "href": "posts/weeknotes-2025-W06.html", + "title": "Week notes (2025-W06)", "section": "", - "text": "The first part of my week largely involved reading and taking notes on methodological texts. I focused on case study design and qualitative coding techniques. I plan to continue my methodology readings on coding techniques, memoing, interview methods and systematic note-taking, as well filling in gaps in my understanding of grounded theory and related debates. These readings are especially useful in this planning stage, but also serve to fill time while I wait for my IRB approval.\nOn that note, I finally submitted my ethics application on Thursday. I expect an expedited review based on the low-risk nature of the work. I posted the materials I submitted on the ethics protocol page.\nI had my biweekly meeting with David, and he was very encouraging.\nI met with Isabel Fortier again yesterday and we came up a list of six projects that may serve as potential cases. We will discuss them in greater depth next week.\nI finally sent some feedback on qc.\nNext week I also need to fulfill a few commitments not as related to the postdoc: I need to work on a peer-review I had committed to, continue assemmbling constructive feedback for qc, and continue going through the DOAJ for the diamond.open-archaeo initiative." + "text": "This week was a bit slower than usual. The highlight was my meeting with Isabel to go over details about harmonization procedures and to discuss projects that may serve as potential cases. I got a better sense of the community composition and the kinds of challenges that are commonly experienced, and Isabel was able to provide me with contacts to get in touch with once I’m ready to begin collecting data.\nI was less active with regards to my methodology notes this week, having spent much time re-organizing and consolidating them. I’m also reading about coding techniques and the foundational principles and debates surrounding grounded theory, but there is a lot of ground to cover and I’m taking it all in before I begin recording my thoughts in a systematic way.\nI did some non-postdoc service this week too. I started a peer-review that I had committed to, participated in the monthly SSLA meeting, and completed the first-pass screening of archaeology-related journals in the Directory of Open Access Journals to verify their inclusion in diamond.open-archaeo." }, { - "objectID": "posts/weeknotes-2025-W03.html", - "href": "posts/weeknotes-2025-W03.html", - "title": "Week notes (2025-W03)", + "objectID": "posts/weeknotes-2025-W04.html", + "href": "posts/weeknotes-2025-W04.html", + "title": "Week notes (2025-W04)", "section": "", - "text": "I’m trying out a new way to track and communicate my progress on this project. Every week I’ll write a post to track the work I’ve been doing and reflect on my activities. I’ll try to maintain this document throughout the week, tidy it up and post it here on Friday afternoons. However the specific process will probably vary as it grows into the rest of my workflow.\nI’m purposefully trying to not tie this into the personal knowledge management trend. It’s for me and my own purposes, and I don’t want to get bogged down with the unabashed managerial phoniness that belies the PKM phenomenon.\nAnyway, I didn’t take notes on my work this past week, but here’s an overview of what I’ve done from memory:\nContinued to take notes on readings produced by the Maelstrom Project and its partners.\nContinued to investigate and maintain notes on potential cases.\nSet up a placeholder document for notes on methodological concerns.\nMet with David for our bi-weekly check-in (meeting notes are private, at least for now). I was also given access to the lab’s private git server but haven’t really had much of a chance to explore what it’s being used for or devise my own plans to make use of it.\nWorked extensively on the ethics protocol. David and I went back and forth deciding on whether this was necessary, given how the project is based on his grant which already has IRB approval. But it’s better to play it safe than sorry, especially when it’s necessary to obtain informed consent. So to this end, I revised the research protocol and responses to the ethics form, and I also drafted an informed consent document. I’ll share all these things once the whole package is put together (probably in a week or two), but my responses to the ethics form already appears on the ethics protocol page.\nI simplified the way private documents are handled in the quarto project and git respository. May still need to so some fiddling, especially for draft blog posts and notes.\nI started drafting a blog post about the potential use of AI/LLMs in my research. Stay tuned.\nOn a related note, I watched this recent video about the use of LLMs in qualitative data analysis, which did not prompt me to draft the post but which is well-timed, nevertheless.\nI worked a bit more on the data management plan, which prompted me to think more about which QDA software I’ll use. I started filling in a university form to use cloud services provided by MaxQDA, but stumbled upon qualitative-coding (abbreviated as qc), an open source CLI-based QDA system. It represents a very innovative approach to QDA rooted in computational thinking and plain text social science, while also remaining true to the core tenets and purpose of QDA, which make it unique and difficult to design software for. If this is the sort of thing that appeals to you, I highly recommend you read the docs.\nI had a bit of trouble installing it and getting it running, but I met remotely with Chris Proctor, who develops the tool through his work at the Computational Literacies Lab, based in the Department of Learning and Instruction at University at Buffalo (SUNY). He helped me resolve some issues, gave me a guided tour of the system and we just talked about the overall state of qualitative data analysis and its tooling. I don’t really have the capacity right now to post everything he showed me but I will definitely be posting about my experiences tinkering around with qc in the coming weeks.\nSimilarly, I asked on Mastodon about whether there are any tools that might support automatic generation of transcripts that include support for specialized notation. A few linguists and conversation analysis scholars responded with recommendations to use GailBot, and with discussion about the tool’s capabilities and limitations. I requested access to the software but haven’t heard back from the dev team yet. I also created a thread on the whisper github repo, which I now realize it a bit of a naive place to put it, and it hasn’t yet gotten any responses.\nI attended a talk from the epidemiology seminar series, which went wayyyy over my head.\nDid my usual amount of engagement on masotodon, I suppose. And I continued to make new friends in the department too :)" + "text": "This week was a bit slower than last. I spent much of it finalizing the content for my IRB application, and the rest preparing for a meeting with a key stakeholder relating to my research.\nThe IRB application is more or less done, just waiting on David and the department head to sign off. It was a major opportunity to re-organize my research protocol and related documents. I shuffled some things over into various placeholder sections in my methodology notes, and pushed a revised research protocol to the website.\nYesterday I posted about on the role of AI in my research. It’s mainly meant to lay out my current state of thinking on AI. I’m not fixed to those ideas, and I think there is much more nuance than I do justice to in that post, but putting it on the page helped me consolidate and put aside some scrambled opinions.\nAfter playing around with qc on Sunday, I started to assemble some feedback. I may post a github issue later this week, once I’ve had a chance to consolidate and edit my thoughts.\nI participated in the weekly CITF logistics update, after which I met with Aklil to discuss the overall strategy for her project and strategize on how we might form the focus groups for that work. We’re gonna meet more regularly, just to share some updates on our respective projects which have a lot in common.\nOn Thursday I met with Isabel Fortier, with the intention of discussing data harmonozation initiatives that might serve as potential cases. It was a bit of a tough meeting. During the first 45 minutes I struggled to communicate the purpose of my work, but I think by the end we reached a greater understanding of what this work will entail and the unique perspective it will bring. One surprising outcome that I still need to think through is Isabel’s suggestion that I slow down a bit, immerse myself more in the world of data harmonization. While she is absolutely right that I’ve been rushing through this first month, I do feel pressure to get the project going and to start accumulating data — I felt a similar pressure when starting my PhD, too. So I need to put my eagerness aside so that the data are relevant and of good enough quality. Isabel offered to schedule regular meetings with me, and even to have me work in the Maelstrom office once a week, and I’m extremely grateful for her support! Plus, I’ll get to have lunch with my mom who works in the same hospital complex, one building over :)" }, { - "objectID": "posts/2024-12-11-technical-specs.html", - "href": "posts/2024-12-11-technical-specs.html", - "title": "Technical specs for this website", + "objectID": "posts/2025-01-24-ai-in-my-work.html", + "href": "posts/2025-01-24-ai-in-my-work.html", + "title": "On the role of AI in my research", "section": "", - "text": "I’m using this website as a way to help organize and share key documents and resources. The research protocols are in flux at this stage in the project’s development, and this will make it easier to distribute up-to-date drafts with partners, while simultaneously enhancing transparency.\nThis post outlines the technical specifications for this website and outlines a roadmap for its further development. It will therefore be continually updated as the site evolves." - }, - { - "objectID": "posts/2024-12-11-technical-specs.html#fundamentals", - "href": "posts/2024-12-11-technical-specs.html#fundamentals", - "title": "Technical specs for this website", - "section": "Fundamentals", - "text": "Fundamentals\nThis website is based on Quarto, a platform for writing and publishing scientific and technical writing. I had used quarto before but without fully understanding it, and now I am starting to see its elegance.\nI had started off using Hugo, but there were too many limitations that Quarto was able to accomodate. You can find an older version of this post reflecting that setup here: #2346852.\nThe site is hosted on GitHub Pages. The repo is located at https://github.com/zackbatist/CITF-Postdoc." - }, - { - "objectID": "posts/2024-12-11-technical-specs.html#generating-pdfs", - "href": "posts/2024-12-11-technical-specs.html#generating-pdfs", - "title": "Technical specs for this website", - "section": "Generating PDFs", - "text": "Generating PDFs\nAs an avid user, one thing I really like about Quarto is the ability to generate PDFs alongside html versions served over the web. I started tinkering with includes but I need to review how Quarto passes info from YAML frontmatter. It is not at all straightforward and I will need to experiment a bit more with this to get the hang of it." - }, - { - "objectID": "posts/2024-12-11-technical-specs.html#archiving-and-version-control", - "href": "posts/2024-12-11-technical-specs.html#archiving-and-version-control", - "title": "Technical specs for this website", - "section": "Archiving and Version Control", - "text": "Archiving and Version Control\nEvery change is tracked using git. I would also like to archive each research protocol in Zenodo once they reach a point of stability. This would ensure that they ca be aassigned DOIs and detailed metadata, which will make them easier to reference.\nHowever, I do not want to rely on Zenodo’s GitHub integration for two reasons: (1) I want this to be as platform-agnostic as possible, and (2) that system relies on GitHub’s release system which operates on the level of the whole repository rather than specific files.\nI might be able to write a custom CI workflow to archive specific files to Zenodo using their API. But, I want to be able to toggle this option, rather than have it occur for every single detected change. Maybe I can accomplish this by pushing the changes that I want to archive to a dedicated branch that the CI workflow is configured to operate on. Or it might be easier to simply do this manually, since I’m not sure I will be using it that often anyway." + "text": "AI is upon is, and although I would probably be ok if it wasn’t around, I have been (and still am, to a certain extent) tempted to use it in my research. So here I’m gonna articulate some of my thoughts on AI. This isn’t written to convince anyone, or even to convince myself. Just to lay out all my thoughts and take stock of my preconceptions, disapointments, hopes and desires, etc.\nAlso, I’m gonna use AI, LLM and whatever other brand names and marketing buzzwords interchangably here. Draw whatever conclusions you want about that.\nI see AI as being potentially useful in a few productive activities I regularly engage in:\n\nTranscribing spoken words into written text\nTranscription is a significant component of processing interview data, and this can be extremely slow work. It’s a lot easier to edit a transcript produced through a computer algorithm rather than start from scratch. I used trint, otter and other similar tools before all the AI hype, and more recently I’ve been using whisper to transcribe voice notes that I record while I’m waiting for the bus or drifting off to sleep. I’m not really sure how they’re much different, to be honest. Is AI just a rebrand of natural language processing in these contexts? Either way, I will most certainly be using some automatic transcrion tool in my research.\nSummarizing, breaking down and simplifying complex bundles of ideas\nI do a lot of reading, and it can be hard to get through everything on my list. I therefore make lots of compromises and refrain from reading some things because I just can’t make enough time to get through everything. I imagine that AI can help summarize some key points across a whole corpus of articles on my to-read pile, and I may try it out once I have time to figure out the right tooling for the job. However, I do gain a lot of value from the process of reading. Specifically, as a scholar of scientific practice, I’m interested in the language and rhetoric authors use to describe and situate their methods and findings, and I’m not sure if automatic summary tools can capture and communicate this nuance in ways that I want.\nGenerating code snippets for data processing and visualization\nThis is arguably the most productive potential application I can imagine. Specifically, I’m thinking about using this to generate R code that processes and visualizies data according to imagined outcomes. This is directly relevant to a project I’m working on where I’ve already finished the workflows for scraping and processing the data, I have the questions I want to ask of it, but I don’t have the practical know-how to generate the code that will allow me to address them. ggplot is just so dense to me, and stitching together code snippets from stack exchange is a major pain in the ass that produces a horrible abomination of code that would not pass the muster of any rigorous code review. What’s more, those queries to search stack exchange are already half-formed AI prompts! At least an AI would generate some harmony in the code, and I might learn something by having a tidy and consistent template.\n\nI’m more ambivalent and critical about using AI in these contexts where it’s been really hyped:\n\nAny form of writing, including generating emails and abstracts\nFor me, writing is a creative process and a way of unerstanding. It’s a mechanism through which I come to learn about something. The experience of drafting and revising a document is crucial to my research process. This is especially important for honing my position as a scholar at the intersection of various disciplinary communities, who have distinct language and modes of communication.\nQuerying for truth claims\nTo be clear, the idea that knowledge can be total, absolute and disembodied is deeply flawed, and the popular reception of AI as a neutral observer and reporter of nature makes me sad. That being said, I’m still ambivalent about the potential for specialized, home-grown LLMs as means of parsing, sorting through and obtaining greater value from under-used resources. There are patterns in even the messiest and least formal documents we create, and even if we can’t draw information from these documents, LLMs may be useful to help us reflect on the circumstances of their creation. I keep thinking about Shawn Graham’s twitter bots in this context (which were not based on AI, but whatever), which attempted to spit out segments of artificial reports and fieldwork drama, which real archaeologists often related and resonded to. These responses were interesting to me, often expressed as collective fascination, titilation or disgust, and reminiscient of the apprehension one might experience when hearing your own voice played back while standing at the opposite end of a long hallway. Reacting to distortions of your own experience from very different perspectives can be a really powerful reflexive exercise.\nAs a brainstorming tool, or as a rubber duck\nI’ve heard about people using AI chatbots as agents to bounce their ideas off of. Kind of like eliza, but for productive work. While I think it’s intriguing, I don’t know where I’d start. Also, drawing up the prompt and figuring out how to ask the right questions may already be enough to get the ideas flowing. I think I already do this in some ways by drafting little ephemeral notes, usually directed toward a specific person or imaginary audience while anticipating their feedback. It also somehow seems like a perverse way to de-socialize work, and in a world where students and postdocs feel increasingly isolated, I’d much rather solicit and provide feedback among peers. This has been the foundation of some of my most solid friendships and professional partnerships, and should be encouraged.\n\nI also have some previously-unstated opinions in relation to some common critiques of AI:\n\nProcess versus product\nAI seems to be really good at devising formulaic outputs. That is, it’s good at getting things to look like things whose shapes are already well-defined. This can be valuable in various use cases, like writing emails according to a template or translating texts between languages. I could imagine it being really helpful for those who are coming into a field where certain skills are taken for granted, such as learning how to write “proper” academic emails as a student who is not fluent in english. Imagine being up against a deadline for a job application, while also being knee-deep in unpaid work to get your name out there; an LLM could be a godsend. So I don’t discount easy outputs as inherently bad. A standard output for one is a week-long struggle for another, so I think this distinction between product and process is a false and misleading dichotomy.\nBad instructions\nSometimes I find it really hard to believe that people could earnestly follow whatever an AI tells them. But I think we’re getting to the point of urban mythmaking, similar to the older wariness about following your GPS into a lake. There’s a story behind every warning sign, even if it’s a projection of what you think might happen if you disregard it.\n“Intelligence”\nOne weird thing about AI branding is the smushing together of some unified idea of what constitutes “intelligence”. We’ve already been through this with “smart” gadgets, which have always just been ways to capture consumer products under a platforms proprietary injected plastic molds and information protocols. AI is literally just a way to sell you a new version of the smart gadget you threw out last year.\nTruthiness, i.e., AI’s ability to sound authoritative while also making false claims\nI cringe at any retort to a screenshot of AI giving a wrong definition of a thing. Accuracy of responses should come secondary to critique of the notion that all forms of knowledge can be presented in terms of absolute, disembodied and universally truths. For example, when people ridicule AI’s inability to identify the capitols of various nation states, I see missed opportunities to challenge the value of any answer that anyone might provide. True subversion would be to reject or re-frame the question and the simplicity with which it is addressed.\nOne another related note, I see a lot of weird parallels between myths about truth claims made by AI and by practitioners of qualitative data analysis (QDA) — and, as a qualitative researcher, this is obviously a bit unsettling. Specifically, in both QDA and AI, there is no actual attempt to make absolute truth claims, but the focus is rather on attempting to identify and draw out meaningful elements of elicitations in a corpus, and to trace patterns between them. In my current opinion, the key difference lies in positionality. Any QDA researcher who laim that their cases are representative of all experiences will be laughed out of the room. Meanwhile, AI is lauded for the claims made by their creators that it can derive unambiguous and concrete knowledge from inherently situated and biased data sources. Humility is key while contributing to collective knowledge bases, and AI risks changing the dynamic away from deriving greater value from constructive discourse and toward a system where the loudest voice in the room wins.\nClimate change\nAI uses a lot of energy, and is therefore said to be wasteful. However I think there are certain wasteful components of AI. For instance, generative models that spit out a full sentence to wrap around the answer to a question don’t have to do all that extra work. Also, not everyone is reliant on fossil fuels, and the critique that AI is necessarily bad for the environment is laden with a thick American accent (as is the case with so many of the loudest opinions on the internet).\nThat being said, there are enormous problems with resource allocation in AI, and I’m not trying to dismiss all concerns. I see these concerns as relating to the distribution of power and wealth in society at large, and AI is one aspect of this. Sometimes I wonder if comparisons can be made between using AI in selective research contexts and eating a burger or a banana, which each have their own environmental costs. But thinking in this way is a bit of a trap.\n\nI also see that rhetoric, including anxieties about AI, differs in the various communities I participate in:\n\nIn digital-x, where x = {archaeology | humanities | librarianship | whatever}\nThere’s a lot of experimentation going on. Honestly, I don’t know much about it and I tend to scroll past any discussion about AI applications in archaeology that appears in my feed. Part of me sees it as a passing trend, but it could be better framed as a wild frontier, as is the case with many other things in digital archaeology. People are still in the process of taming the landscape, to make it better suit their needs, and maybe I’ll join in once the settlement is established. But I’m not personally motivated by the dynamism of the current landscape, at least in this particular domain.\nEpidemiology, biostats, public health\nI’m still too new in this community to really make sense of this yet. I’ll continue to watch and learn and listen.\nBroader social science and humanities, as well as libraries, archives and museums\nCritique tends to follow broader, more abstract, and more common-sense lines of thought. In my view, much of this does not really account for the material problems and imperfections in which the social sciences and humanities operate. AI is a lifeline for many people in an overworked, overburdened, under-resourced and hyper-competitive environment, and tut-tutting around how other people use AI sometimes comes across as tone-deaf and disrespectful. Some criticisms of AI being used in real, practical circumstances make me second guess critics’ supposed commitments to improving the social experience of research. The fundamental problem is inequitable access to financial and material resources, and AI’s prevalence is a major symptom of, or — depending on your perspective — resolution to that. People’s who recognize this have no choice but to post broader and more abstract criticisms, which come across as somewhat hollow when disconnected from real and tangible experiences.\nSenior faculty\nProbably the most ambivalent of all communities are senior faculty, who want AI to be useful and will test the waters without fully committing. Which is fine and very prudent, and honestly I identify most with this perspective, despite my position as a lowly postdoc.\nGrad students\nI engage with many grad students. I share my workspace with grad students and encounter them constantly in my day to day neighbourhood forays, where I overhear and sometimes participate in conversations about AI. In my new work environment (Epidemiology, Biostatistics and Occupational Health), the grad students who I engage with have a relatively positive perception of AI. They seem to find greater value in the ability to automate complex processes, using it as a black box of sorts, with predictable and abstracted inputs and outputs, which they see as especially helpful for coding. Outside of this space I’m encountering way more diversity of thought on AI, and I’m not quite sure how to group these viewpoints to structure a proper reaction. I think this in fact contributes to the multitude of perspectives, since no one really cares that much one way or the other to really have a strong opinion (though I sense an overwhelming dissatisfaction when it comes to AI in consumer contexts; this post is largely about productive uses of AI in research and pedagogy).\nI was also told about students learning RStats by just having AI generate their code. The person who pointed this out to me related this to the growing misconception that to learn stats you first need to learn how to code. This in turn relates to the sense that to learn how to do RStats, you just need to memorize a series of steps and copy the text from the slides into the IDE. So, in the end, AI reveals the inadequacy of the teaching mechanisms for programming and stats classes, similarly to how AI has revealed the inadequacy of essay-writing as a pedagogical technique.\nOn the other hand, some students are concerned about dulling their skills, or even not being able to take advantage of opportunities to learn new skills, due to the temptation to automate these tasks. Some upper-year PhD students are glad that they were trained in the fundamentals prior to the AI hype wave. This makes me wonder how students are determining what skills they think they need to know how to do on their own and what is worth running through an LLM. Does it basically operate as a bullshit sensor, where you can smell from a distance that the work is just gonna be tedium and irrelevant? Or is it more out of practical necessity, where you’re stretched so thin that you simply have to rely on these tools to achieve anything meaningful, almost as a mechanism for salvaging one’s work from the claws of austerity? In either case, this points to PhD programs’ inadequacy to match students’ needs and desires, and overwhelming amount of administravia or (seemingly) irrelevant work that students are made to do, which get in the way of their true interests.\n\nMaybe I’ll have more to share some other time." }, { - "objectID": "posts/2024-12-09-first-team-meeting.html", - "href": "posts/2024-12-09-first-team-meeting.html", - "title": "Reflection on first team meeting", + "objectID": "posts/2024-12-09-hello-world.html", + "href": "posts/2024-12-09-hello-world.html", + "title": "Hello World!", "section": "", - "text": "Last week (2024/12/04) I finally met with David Buckeridge, Tanya Murphy and Aklil Noza in person. The meeting was meant to convey my vision for the project to the whole team, to align perspectives, and to articulate how this will actually work in practice.\nThe gist is that I will be investigating the role of social and cultural factors in data-sharing initiatives such as CITF and other Maelstrom-affiliated projects, and how these relate to, overlap with, or conflict with technical and institutional/administrative factors. To be clear, these are all very important aspects of data-sharing, but we generally recognized that the social and cultural aspects are under-explored relative to their impact.\nWe briefly talked about how we will go about selecting cases, and I emphasized the importance of strategic case selection. This also involves carefully articulating the project’s goals so that the cases will complement them. We agreed that the dataset will likely comprise between 12-15 interviews of around 60-90 minutes in length with representatives from 4-5 cases (one of them being CITF), in addition to representatives of the Maelstrom team. Maelstrom will serve as a “fixed point” that limits the scope of the cases’ breadth and ensures that participants have a common frame of reference. It also potentially allows me to “offload” or “consolidate” reference to technical and administrative aspects of data-sharing through targeted interviews with Maelstrom personnel, instead of dealing with those things with the representatives for each case.\nWe discussed timelines and overlap with Aklil’s work, which will be more concerned with focus groups with CITF databank users. There is definitely overlap with the emphasis of my own work and we will coordinate data collection to enhance the potential for analytical alignment.\nAfter the meeting I chatted with Tanya and Aklil who helped familiarize me with the bigger-picture theoretical discourse and tensions in epidemiology. Much of it seemed familiar since these concerns are common across disciplines, but I still need to read more to concretize my understanding. Tanya recommended I read the “Baby Rothman” which is a condensed version of a very long-lived textbook in this field, among a few other papers she sent me.\nOverall, this meeting got me really excited about this project :)" + "text": "Welcome to the website for my CITF Postdoc! This will serve as a hub for documenting and sharing my work. I decided to do this as a way of managing and sharing always-updated drafts of research protocols with my supervisor and team members, but it is also generally useful for keeping my thoughts organized. I will also use this blog section to write my thoughts as the project progresses." }, { - "objectID": "notes/potential-cases.html", - "href": "notes/potential-cases.html", - "title": "Potential cases", + "objectID": "posts.html", + "href": "posts.html", + "title": "Blog", "section": "", - "text": "Isabel Fortier came up with a shortlist based on consultations to help determine which Maelstrom partner projects may serve as potential cases. We then met on 2025-02-04, when, among other topics, we discussed the shortlist.\nSee the case selection protocol for further details on the parameters that guide how cases are to be determined." + "text": "Date\n\n\nTitle\n\n\nCategories\n\n\n\n\n\n\nFeb 21, 2025\n\n\nWeek notes (2025-W08)\n\n\nweek notes\n\n\n\n\nFeb 14, 2025\n\n\nWeek notes (2025-W07)\n\n\nweek notes\n\n\n\n\nFeb 7, 2025\n\n\nWeek notes (2025-W06)\n\n\nweek notes\n\n\n\n\nJan 31, 2025\n\n\nWeek notes (2025-W05)\n\n\nweek notes\n\n\n\n\nJan 25, 2025\n\n\nWeek notes (2025-W04)\n\n\nweek notes\n\n\n\n\nJan 24, 2025\n\n\nOn the role of AI in my research\n\n\nAI / LLM, Methods, QDA\n\n\n\n\nJan 18, 2025\n\n\nWeek notes (2025-W03)\n\n\nweek notes\n\n\n\n\nDec 18, 2024\n\n\nTechnical specs for this website\n\n\nwebsite\n\n\n\n\nDec 9, 2024\n\n\nReflection on first team meeting\n\n\nmeeting notes, general thoughts\n\n\n\n\nDec 9, 2024\n\n\nHello World!\n\n\nintroduction, website\n\n\n\n\n\nNo matching items", + "crumbs": [ + "Blog" + ] }, { - "objectID": "notes/potential-cases.html#general-notes", - "href": "notes/potential-cases.html#general-notes", - "title": "Potential cases", + "objectID": "notes/methodology-notes.html", + "href": "notes/methodology-notes.html", + "title": "Methodology notes", "section": "", - "text": "Isabel Fortier came up with a shortlist based on consultations to help determine which Maelstrom partner projects may serve as potential cases. We then met on 2025-02-04, when, among other topics, we discussed the shortlist.\nSee the case selection protocol for further details on the parameters that guide how cases are to be determined." + "text": "This document is an overview of methodological topics and concerns. It is a place where I think through and justify my methodological decisions, and identify the methods and procedures through which I implement them." }, { - "objectID": "notes/potential-cases.html#possible-candidates", - "href": "notes/potential-cases.html#possible-candidates", - "title": "Potential cases", - "section": "Possible candidates", - "text": "Possible candidates\n\nCITF\n\n\nReACH\nTitle: Stress and Anxiety During Pregnancy and Length of Gestation Harmonization Initiative + ReACH\nContact: Julie Bergeron \nReason: A small project and a very complex infrastructure-oriented network coordinated by Julie Bergeron. Both projects are finalized.\nNotes: Julie Bergeron was a PhD student at Maelstrom, and Isabel says that she is probably the most knowledgeable person regarding data harmonization who I will encounter during my research. She worked on her dissertation project (Stress and Anxiety During Pregnancy and Length of Gestation Harmonization Initiative) while also leading work on ReACH, and her dissertation essentially served as one of a few pilot projects under the aegis of ReACH. ReACH was led by Isabel as its PI.\nBoth projects are complete, but Isabel thinks that Julie Bergeron will be able to share some significant insights on this past work. My instinct is that this presents an opportunity to explore how/whether harmonization is suited for doctoral training, the role of pilot projects within broader initiatives, and impact that closeness to the method of data harmonization might have.\nLinks:\n\nhttps://www.maelstrom-research.org/network/reach\n\n\n\nCAPACIty\nTtle: Capacity: Building CAPACIty for pediatric diabetes research and quality improvement across Canada\nAcronym: CAnadian PediAtric diabetes ConsortIum (CAPACIty)\nContact: Shazhan Amed \nReason: Across Canada, focus on clinical data. A lot of work already achieved, and harmonization will start soon. Will use a federated approach for statistical analysis.\nNotes: A network of 15 childhood diabetes centers from across Canada. Went through four years of administrative work, and is now just starting harmonization after finally going through all theose hurldes, despite being very eager to get into the data work early on. Despite these challenges, Isabel thinks they will be very receptive to participating in the study.\n\n\nSHAIRE\nTitle: SHAIRE: Scientific & Health Data Assets In Respiratory Research\nContact: Sanja Stanojevic \nReason: New project just starting, very interesting and dynamic.\nNotes: Extremely new, just got funding very recently. I don’t know that much, to be honest. Could potentially provide redundant value to my study as Capacity, but need to find out more.\n\n\nRespiratory study\nTitle: High-dose docosahexaenoic acid for bronchopulmonary dysplasia severity in very preterm infants: A collaborative individual participant data meta-analysis\nContact: Isabelle Marc \nReason: Very specific and small project, very clinical.\nNotes: A very small project, harmonizing two datasets. I asked if this scale of work is common and Isabel says that it is, so it’s not an outlier.\nLinks:\n\nhttps://maelstrom-research.org/study/n3pi-hi\n\n\n\nMORGAM\nTitle: MOnica Risk, Genetics, Archiving and Monograph\nContact: Kari Kuulasmaa \nReason: European, started several years ago.\nNotes: Older project, ended around 10 years ago, the PI is retired. Might be valuable for looking at broader impact and potential offshoots after the work has been completed.\nLinks:\n\nhttps://www.maelstrom-research.org/network/morgam\n\n\n\nLifecycle\nVery improvized approach to data harmonization, did a lot of “manual” work. According to Isabel, Julie Bergeron will be able to tell me more.\nLinks:\n\nhttps://pmc.ncbi.nlm.nih.gov/articles/PMC7387322" + "objectID": "notes/methodology-notes.html#multicase-studies", + "href": "notes/methodology-notes.html#multicase-studies", + "title": "Methodology notes", + "section": "Multicase Studies", + "text": "Multicase Studies\nThese notes describe the features, affordances and limitations of case study research, and articules factors correspoding with variable kinds of case studies.\nI do notice a distinction between two schools of thought, which seem to be spearheaded by Stake and Yin. I generally favour Stake’s flexible approach, and it seems well aligned with other methodological works I’ve been reading (e.g. Abbott 2004; Ragin and Becker 1992).\n\nStake’s Approach\nIn case-study research, cases represent discrete instances of a phenomenon that inform the researcher about it. The cases are not the subjects of inquiry, and instead represent unique sets of circumstances that frame or contextualize the phenomenon of interest (Stake 2006: 4-7).\nCases usually share common reference to the overall research themes, but exhibit variations that enable a researcher to capture different outlooks or perspectives on matters of common concern. Drawing from multiple cases thus enables comprehensive coverage of a broad topic that no single case may cover on its own (Stake 2006: 23). In other words, cases are contexts that ascribe particular local flavours to the activities I trace, and which I must consider to account fully for the range of motivations, circumstances and affordances that back decisions to perform activities and to implement them in specific ways.\nMoreover, the power of case study research derives from identifying consistencies that relate cases to each other, while simultaneously highlighting how their unique and distinguishing facets contribute to their representativeness of the underlying phenomon. Case study research therefore plays on the tensions that challenge relationships among cases and the phenomenon that they are being called upon to represent (Ragin 1999: 1139-1140).\nStake (2006: 4-6) uses the term quintain1 to describe the group, category or phenomenon that bind together a collection of cases. A quintain is an object, phenomenon or condition to be studied – “a target, not a bull’s eye” (Stake 2006: 6). “The quintain is the arena or holding company or umbrella for the cases we will study” (Stake 2006: 6). The quintain is the starting point for multi-case research.\nAccording to Stake (2006: 6):\n\nMulticase research starts with the quintain. To understand it better, we study some of its single cases — its sites or manifestations. But it is the quintain we seek to understand. We study what is similar and different about the cases in order to understand the quintain better.\n\nStake (2006: 8) then goes on:\n\nWhen the purpose of a case is to go beyond the case, we call it an “instrumental” case study When the main and enduring interest is in the case itself, we call it “intrinsic” case study (Stake 1988). With multicase study and its strong interest in the quintain, the interest in the cases will be primarily instrumental.\n\nAbbott’s (2004: 22) characaterization of Small-N comparison is very reminiscient of Stake’s (2006) account of the case-quintain dialectic:\n\nSmall-N comparison attempts to combine the advantages of single-case analysis with those of multicase analysis, at the same time trying to avoid the disadantages of each. On the one hand, it retains much information about each case. On the other, it compares the different cases to test arguments in ways that are impossible with a single case. By making these detailed comparisons, it tries to avoid the standard critcism of single-case analysis — that one can’t generalize from a single case — as well as the standard criticism of multicase analysis — that it oversimplifies and changes the meaning of variables by removing them from their context.\n\nIt should be noted that case study research limits my ability to define causal relationships or to derive findings that may be generalized across the whole field of epidemiology. This being said, case study research allows me to articulate the series of inter-woven factors that impact how epidedemiological researchers coordinate and participate in data-sharing initiatives, while explicitly accounting for and drawing from the unique and situational contexts that frame each case.\nStake (2006: 23) recommends selecting between 4-10 cases and identifies three main criteria for selecting cases:\n\nIs the case relevant to the quintain?\nDo the cases provide diversity across contexts?\nDo the cases provide good opportunities to learn about complexity and contexts?\n\n\nFor qualitative fieldwork, we will usually draw a purposive sample of cases, a sample tailored to our study; this will build in variety and create opportunities for intensive study (Stake 2006: 24).2\n\nStake’s (2010: 122) prioritizes doing research to understand something or to improve something, and I generally agree with his rationalization; research helps reframe problems and establish different decision options.\n\n\nYin’s Approach\nAccording to Yin (2014: 16), “a case study is an empirical inquiry that investigates a contemporary phenomenon (the ‘case’) in depth and within its real-world context, especially when the boundaries between phenomenon and context may not be clearly evident.”\nHe goes on to document some features of a case study: “A case study inquiry copes with the technically distinctive situation in which there will be many more variables of interest than data points, and as one result relies on multiple sources of evidence, with data needing to converge in a triangulating fashion, and as another result benefits from the prior development of theoretical propositions to guide data collection and analysis.” (Yin 2014: 17)\nYin (2014) is more oriented toward what he refers to as a realist perspective, which he pits against relativist and interpretivist perspectives (used interchangably, it seems), and which I might refer to as constructivist. He characterizes relativist perspectives as “acknowledging multiple realities having multiple meanings, with findings that are observer dependent”. His prioriting of a realist approach corresponds with the analysis by Yazan (2015), who compared Yin with Stake and Merriam. According to Yazan (2015: 137), Yin evades making statements about his epistemic commitments, and is characterized as post-positivist.\nYin (2014) is very concerned with research design in case study research He posits that, in a colloquial sense, “a research design is a logical plan for getting from here to there, where here may be defined as the initial set of questions to be answered, and there is some set of conclusions (answers) about these questions.” (Yin 2014: 28)\nYin distinguishes between a research design and a work plan. A research design deals with a logical problem, whereas a work plan deals with a logistical problem. Seems reminiscient of Brian Cantwell Smith’s distinction between skeletons and outlines.\nYin lists five components of a research design:\n\nA case study’s questions;\nits propositions, if any;\nits unit(s) of analysis;\nthe logic linking the data to the propositions; and\nthe criteria for interpreting the findings.\n\nInterestingly, I have been instinctively following these steps, and am currently hovering somewhere between components 3 and 4, while dipping back to 2 once in a while too.\nThe problem of defining the unit of analysis is salient to me right now. According to Yin (2014: 32), the unit of analysis may change as the project progresses, depending on initial misconceptions (he uses the example of a unit of analysis changing from neighbourhoods to small groups, as contextualized by the socio-geographical entity of the neighbourhood, which is laden with issues of class, race, etc). In my own situation, the unit of analysis may hover between the harmonization initiative, the people, activities or infrastructures that make it work.\nIn the section on criteria for interpreting the findings, Yin emphasizes the role of rival theories, which is akin to a concern with falsifiability as a means of validating truth claims, and which betrays his positivist leanings. This may be compared with Stake’s emphasis on triangulation, which is more concerned with internal cohesiveness. Similarly, Yin cites Corbin and Strauss regarding the role of theory or theoretical propositions in research design, which similarly reveals a concern with rigorous upfront planning and strict adherence to research design as a key aspect of deriving valid findings.\nRegarding generalizability, Yin (2014: 40-41) states that “Rather than thinking about your case as a sample, you should think of it as the opportunity to shed empirical light about some theoretical concepts or principles, not unlike the motive of a laboratory investigator in conceiving of and then conducting a new experiment.” He goes on to state that case studies tend to strive for analytic generalizations that go beyond the specific case that has been studied, and which apply to other concrete situations rather than just abstract theory building.\n\n\nLogistics of case study design\n\nPreparing to select case study data\nYin (2014: 72-73) identifies five desired attributes for collecting case study data:\n\nAsk good questions — and interpret answers fairly.\n\n“As you collect case study evidence, you must quickly review the evidence and continually ask yourself why events or perceptions appear as they do.” (73)\nA good indicator of having asked good questions is mental and emotional exhaustion at the end of each fieldwork day, due to the depletion of “analytic energy” associated with being attention on your toes. (73-74)\n\nBe a good “listener” not trapped by existing ideologies or preconceptions.\n\nSensing through multiple modalities, not just spoken words.\nAlso subtext, as elicited through choices of terms used, mood and affective components. (74)\n\nStay adaptive, so that newly encountered situations can be seen as opportunities, not threats.\n\nRemember the original purpose but willing to adapt to unanticipated circumnstances. (74)\nEmphasize balancing adaptability with rigour, but not with rigidity. (75)\n\nHave a firm grasp of what is being studied, even when in an exploratory mode.\n\nNeed to do more than merely record data, but interpret information as they are being collected and to know immedately whether there are contradictions or complementary statements to follow-up on. (75-76)\n\nAvoid biases of being sensitive to contrary evidence, also knowing how to conduct research ethically.\n\nMaintain strong professional competence, including keeping up with related research, ensuring accuracy, striving for credibility, and knowledging and mitigating against bias.\n\n\nYin advocates for adoption of case study protocols. He provides an example of a table of contents for case study protocols, which generally comprise four sections:\n\nOverview of the case study\nData collection procedures\nData collection questions\nGuide for the case study report\n\n\n\nTriangulation\nTriangulation is a process of gaining assurance. Also sometimes called crystallization.\n“Each important finding needs to have at least three (often more) confirmations and assurances that key meanings are not being overlooked.” (Stake 2006: 33) Triangulation is a process of repetitous data gathering and critical review of what is being said. (Stake 2006: 34)\nWhat needs triangulation? (Stake 2006: 35-36)\n\nIf the description is trivial or beyond question, there is no need to triangulate.\nIf the description is relevant and debatable, there is much need to triangulate.\nIf the data are critical to a main assertion, there is much need to triangulate.\nIf the data are evidence for a controversial finding, there is much need to triangulate.\nIf a statement is clearly a speaker’s interpretation, there is little need to triangulate the quotation but not its content.\n\nStake (2006: 37) cites Denzin (1989) who highlighted several kinds of triangulation, leading to a few advisories:\n\nFind ways to use multiple rather than single observers of the same thing.\nUse second and third perspectives, i.e. the views of teachers, student and parents.\nUse more than one research method on the same thing, i.e. document review and interview.\nCheck carefully to decide how much the total description warrants generalization.\n\nDo your conclusions generalize across other times or places?\nDo your conclusions about the aggregate generalize to individuals?\nDo findings of the interaction among individuals in one group pertain to other groups?\nDo findings of the aggregate of these people generalized to a population?\n\n\n\n\nCross-Case Analysis Procedure\nStake (2006: Chapter 3) lays out a procedure for deriving synthetic findings from data collected across cases. He frames this in terms of a dialectic between cases and quintains. He identifies three tracks (Stake 2006: 46):\n\nTrack 1: Maintains the case findings and the situationality.\nTrack 2: Merges similar findings, maintaining a little of the situationality.\nTrack 3: The most quanitative track, shifts the focus from findings to factors.\n\nAccording to Stake, case reports should be created independently and then brought together by a single individual when working in a collaborative project. In keeping with the case-quintain dialectic, this integration must involve strategically putting the cases aside and bringing them back in to identify convergences and divergences, similarities and differences, normalitities and discrepancies among them.\nThere is some detailed discussion about different kinds of statements, i.e. themes, findings, factors and assertions, but I find this a bit too much detail for me to get at at this point in mymethodological planning. In general though, Stake documents a process whereby an analyst navigates back and forth between the general and the situational, presenting tentativr statements that are shored up, modified or discarded through testing compatability of the evidence across cases.\n\n\nSingle cases\nStake (2000) is concerned with identifying what can be learned from a single case. He (2000: 437) identifies three kinds of cases:\n\nIntrinsic case studies as being driven by a desire to understand the particular case.\nInstrumental case studies are examined “mainly to provide insight into an issue or to redraw a generalization.”\nCollective case studies “investigate a phenomenon, population or general condition”.\n\nStake (2000) frames case research around a tension between the particular and the general, which echoes the case-quintain dilemma he described in (Stake 2006: 4-6).\n\n\nSome scattered practical guidance\nStake (2006: 18-22) provides a detailed and realistic overview of common challenges involved in collaborative qualitative research. This could be handy in future work when planning a multicase project involving multiple researchers.\nStake (2006: 29-33) provides guidance on how to plan and conduct interviews in multicase research, including a series of helpful prompts and questions to ask yourself while designing the interview. One thing that stands out is his recommendation that an interview should be more about the interviewee than about the case. It’s necessary to find out about the interviewee to understand their interpretations, but what they reveal about the quintain is more important.\nOn page 34, Stake (2006) also provides some practical tips for documenting and storing data, after Huberman and Miles (1994).\nStake (2006: Chapter 4) includes a chapter on procedures for reporting the findings, and I may return to this later on once I need to initiative this phase of work. It addresses concerns about how to articulate comparisons, concerns about generalization, and how to handle advocacy based on findings.\nSee Stake (2006) Chapter 5 for a step-by-step overview of a multicase study analysis. The rest of the volume after that includes three very detailed examples from his own work." }, { - "objectID": "notes/potential-cases.html#recommended-but-no-longer-considered", - "href": "notes/potential-cases.html#recommended-but-no-longer-considered", - "title": "Potential cases", - "section": "Recommended but no longer considered", - "text": "Recommended but no longer considered\n\nCanPath\nTitle: CanPath, formerly called the Canadian Partnership for Tomorrow’s Health\nContact: Noah Frank \nReason: One of the most important harmonization initiative, infrastructure oriented, long-term started more than 10 years ago.\nNotes: Other contacts include John Mclaughlin and Paula Robson.\nMy instinct is to look at how things change over the course of a very long and institutional initiative, especially across discrete phases marked by leadership transitions. But the history here is so vast and I will probably not get much of it through a series of interviews.\n\n\nMindmap\nPromoting mental well-being and healthy ageing in cities. Seems very similar to CanPath in terms of scope and governance, and I would likely face similar challenges." + "objectID": "notes/methodology-notes.html#grounded-theory", + "href": "notes/methodology-notes.html#grounded-theory", + "title": "Methodology notes", + "section": "Grounded theory", + "text": "Grounded theory\nThese notes are largely drawn from Kathy Charmaz (2000), which I understand to be a fairly balanced and comprehensive overview of the Glaser / Strauss and Corbin debate, and of the situation of specific methods and techniques in relation to these different stances. I also value Charmaz’s position as someone who subscribes to her constructivist approach.\nAccording to Charmaz(2000: 509):\n\nEssentially, grounded theory methods consist of systematic inductive guidelines for collecting and analyzing data to build middle-range theoretical frameworks that explain the collected data.\n\nCharmaz(2000: 511) goes on to situate grounded theory in relation to what was the norm prior to its invention:\n\nGlaser and Strauss’s (1967) work was revolutionary because it challenged (a) arbitrary divisions between theory and research, (b) views of qualitative research as primarily a precursor to more “rigorous” quantitative methods, (c) claims that the quest for rigor made qualitative research illegitimate, (d) beliefs that qualitative methods are impressionistic and unsystematic, (e) separation of data collection and analysis, and (f) assumptions that qualitative research could produce only descriptive case studies rather than theory development (Kathy Charmaz 1995).\n\nPrior to Glaser and Strauss (1967), qualitative analysis was taught rather informally — they led the way in providing written guidelines for systematic qualitative data analysis with explicit procedures for data analysis (Kathy Charmaz 2000: 512)\nGlaser brought his very positivist assumptions from his work at Columbia, and Strauss’ work in Chicago with Herbert Blumer and Robert Park infused a pragmatic philosophical approach to the study of process, action and meaning that reflects symbolic interactionism.\n\nGlaser\nGlaser’s position comes close to traditional positivism, with assumptions of an objective, external reality and a neutral observer who discovers data. and a reductionist form of inquiry of manageable research problems. According to Charmaz (2000: 511), regarding Glaser’s approach:\n\nTheoretical categories must be developed from analysis of the collected data and must fit them; these categories must explain the data they subsume. This grounded theorists cannot shop their disciplinary stores for preconceived concepts and dress their data in them. Any existing concept must earn its way into the analysis. … The relevance of a grounded theory derives from its offering analytic explanations of actual problems and basic processes in the research setting. A grounded theory is durable because researchers can modify their emerging or established analyses as conditions change or further data are collected.\n\n\n\nCorbin and Strauss\nStrauss and Corbin assume an objective reality, aim toward unbiased data collection, propose a series of technical procedures, and espouses verification. However, they are postpositivism because they propose giving voice to their respondents,3 representing them as accurately as possible, discovering and reckoning with how their respodents’ views on reality differ from their own, and reflecting on the research process as one way of knowing.\nCorbin and Strauss (1990) “gained readers but lost the sense of emergence and open-ended character of Strauss’s earlier volume and much of his empirical work. The improved and more accessible second edition of Basics (Strauss and Corbin 1998) reads as less prescriptive and aims to lead readers to a new way of thinking about their research and about the world.” (Kathy Charmaz 2000: 512)\nStrauss apparently became more insistent that grounded theory should be more verificational in nature in personal communications.\nGlaser (1992) responded to Strauss and Corbin (1990), repudiating what he perceived as forcing preconceived questions and frameworks on the data. Glaser considered it better to allow theory to “emerge” from the data, i.e. to let the data speak for themselves.\nCharmaz identifies these two approaches as having a lot in common: hey both advocate for mitigating factors that would hinder objectivity and minimize intrusion of the researcher’s subjectivity, and they are both embedded in positivist attitudes, with a researcher sitting outside the observed reality; Glaser exemplifies these through discovering and coding data, and using systematic comparative methods, whereas Strauss and Corbin maintain a similar distance through their analytical questions, hypotheses and methodological applications. They both engage in “silent authorship” and usually write about their data as distant experts (Kathy Charmaz and Mitchell 1996).\n\n\nConstuctivist Grounded Theory\n\nConstructivist grounded celebrates firsthand knowledge of empirical worlds, takes a middle ground between postmodernsm and positivism, and offers accessible methods for taking qualitative research into the 21st century. (510)\n\n\nThe power of grounded theory lies in its tools for understanding empirical worlds. We can reclaim these tools from their positivist underpinnings to form a revised, more open-ended practice of grounded theory that stresses its emergent, constructivist elements. We can use grounded theory methods as flexible, heuristic strategies rather than as formulaic procedures. (510)\n\nThree aspects to Charmaz’s argument (510):4\n\nGrounded theory strategies need not be rigid or prescriptive;\na focus on meaning while using grounded theory furthers, rather than limits, interpretive understanding; and\nwe can adopt grounded theory strategies without embracing the positivist leanings of earlier proponents of grounded theory.\n\nRepudiation of the notion that data speak for themselves, that data do not lie. Recognition that data are constructs of the rsearch process, are framed by the questions we ask informants and the methodological tools of our collection procedures.\nCharmaz (2000: 515) advocates for what seems to be a dialogical approach to coding, between researcher and the data:\n\nWe should interact with our data and pose questions to them while coding. Coding helps us to gain a new perspective on our material and to focus further data collection, and may lead us in unforeseen directions. Unline quantitative research that requires data to fit into preconceived standardized codes, the researcher’s interpretations of data shape his or her emergent codes in grounded theory.\n\nDistinguishes articulates open/initial coding as proceeding line by line to get a general sense of what the data contains. It is meant to keep the researcher close to the data, to remain attuned to the subjects’ views of their realities.\n\nLine-by-line coding sharpens our use of sensitizing concepts — that is, those background ideas that inform the overall research problem. Sensitizing concepts offer eays of seeing, organizing, and understanding experience; they are embedded in our disciplinary emphases and perspectival proclivities. Although sensitizing conceots may deepen perception, they provide starting points for building analysis, not ending points for evading it. We may use sensitizing concepts only as points of departure from which to study the data.\n\nMuch of the rest of the Kathy Charmaz (2000) paper is an overview of coding and memoing methods, as well as theoretical sampling. The emphasis is on situating these techniques in the Glaser / Strauss and Corbin debate, and it will be better to refer to Kathy Charmaz (2014) for in-depth notes on these techniques.\nCharmaz (2000: 521-522) provides an apt account of a significant critique of grounded theory, and poses her constructivist approach as a potential means of resolving it. Specifically, she refers to the notion that grounded theory (as traditionally conceived by both Glaser and Strauss and Corbin) “fractures” the data, making them easier to digest in an analytical sense, but also making it more difficult to engage with in a holistic manner. This is precisely the point of the original approach, to present qualitative data as data — as conceived and valued by quantitative researchers, i.e. as discrete, corpuscular, disembodied, re-arrangable and distant entities. The text of these two large paragraphs is copied here:\n\nConrad (1990) and Riessman (1990) suggest that “fracturing the data” in grounded theory research might limit understanding because grounded theorists aim for analysis rather than the portrayal of subjects’ experience in its fullness. From a grounded theory perspective, fracturing the data means creating codes and categories as the researcher defines themes within the data. Glaser and Strauss (1967) propose this strategy for several reasons: (a) to help the researcher avoid remaining immersed in anecdotes and stories, and subsequently unconsciously adopting subjects’ perspectives; (b) to prevent the researcher’s becoming immobilized and overwhelmed by voluminous data; and (c) to create a way for the researcher to organize and interpret data. However, criticisms of fracturing the data imply that grounded theory methods lead to separating the experience from the experiencing subject, the meaning from the story, and the viewer from the viewed. In short, the criticisms assume that the grounded theory method (a) limits entry into subjects’ worlds, and thus reduces understanding of their experience; (b) curtails representation of both the social world and subjective experience; (c) relies upon the viewer’s authority as expert observer; and (d) posits a set of objectivist procedures on which the analysis rests.\nResearchers can use grounded theory methods to further their knowledge of subjective experience and to expand its representation while neither remaining external from it nor accepting objectivist assumptions and procedures. A constructivist grounded theory assumes that people create and maintain meaningful worlds through dialectical processes of conferring meaning on their realities and acting within them (Bury 1986; Mishler 1981). Thus social reality does not exist independent of human action. Certainly, my approach contrasts with a number of grounded theory studies, methodological statements, and research texts (see, e.g., Chenitz and Swanson 1986; Glaser 1992; Martin and Turner 1986; Strauss and Corbin 1990; Turner 1981). By adopting a constructivist grounded theory approach, the researcher can move grounded theory methods further into the realm of interpretive social science consistent with a Blumerian (1969) emphasis on meaning, without assuming the existence of a unidimensional external reality. A constructivist grounded theory recognizes the interactive nature of both data collection and analysis, resolves recent criticisms of the method, and reconciles positivist assumptions and postmodernist critiques. Moreover, a constructivist grounded theory fosters the development of qualitative traditions through the study of experience from the standpoint of those who live it.\n\nCharmaz’s (2000: 523) proposal for a re-visioned grounded theory poses research as a materializing process:\n\nA re-visioned grounded theory must take epistemological questions into account. Grounded theory can provide a path for researchers who want to continue to develop qualitative traditions without adopting the positivistic trappings of objectivism and universality. Hence the further development of a constructivist grounded theory can bridge past positivism and a revised future form of interpretive inquiry. A revised grounded theory preserves realism through gritty, empirical inquiry and sheds positivistic proclivities by becoming increasingly interpretive.\n\nCharmaz (2000: 523) addresses realism and truth in constructivist grounded theory, and explicitly relates it to Blumerian situated interactionism:\n\nA constructivist grounded theory distinguishes between the real and the true. The constructivist approach does not seek truth — single, universal, and lasting. Still, it remains realist because it addresses human realities and assumes the existence of real worlds. However, neither human realities nor real worlds are unidimensional. We act within and upon our realities and worlds and thus develop dialectical relations among what we do, think, and feel. The constructivist approach assumes that what we take as real, as objective knowledge and truth, is based upon our perspective (Schwandt 1994). The pragmatist underpinnings in symbolic interactionism emerge here. Thomas and Thomas (1928: 572) proclaim, “If human beings define their situations as real, they are real in their consequences”. Following their theorem, we must try to find what research participants define as real and where their definitions of reality take them. The constructivist approach also fosters our self-consciousness about what we attribute to our subjects and how, when, and why researchers portray these definitions as real. Thus the research products do not constitute the reality of the respondents’ reality. Rather, each is a rendering, one interpretation among multiple interpretations, of a shared or individual reality. That interpretation is objectivist only to the extent that it seeks to construct analyses that show how respondents and the social scientists who study them construct those realities — without viewing those realities as unidimensional, universal, and immutable. Researchers’ attention to detail in the constructivist approach sensitizes them to multiple realities and the multiple viewpoints within them; it does not represent a quest to capture a single reality.\nThus we can recast the obdurate character of social life that Blumer (1969) talks about. In doing so, we change our conception of it from a real world to be discovered, tracked, and categorized to a world made real in the minds and through the words and actions of its members. Thus the grounded theorist constructs an image of a reality, not the reality — that is, objective, true, and external.\n\nOn the other hand, Charmaz (2000: 524) frames objectivist grounded theory as believing in some kind of truth:\n\nObjectivist grounded theory accepts the positivistic assumption of an external world that can be described, analyzed, explained, and predicted: truth, but with a small t. That is, objectivist grounded theory is modifiable as conditions change. It assumes that different observers will discover this world and describe it in similar ways That’s correct — to the extent that subjects have comparable experiences (e.g., people with different chronic illnesses may experience uncertainty, intrusive regimens, medical dominance) and viewers bring similar que-tions, perspectives, methods, and, subsequently, concepts to analyze those experiences. Objectivist grounded theorists often share assumptions with their research participants — particularly the professional participants. Perhaps more likely, they assume that respondents share their meanings. For example, Strauss and Corbin’s (1990) discussion of independence and dependence assumes that these terms hold the same meanings for patients as for researchers.\n\nCharmaz (2000: 525) further embeds construvist grounded theory as a way to fulfill Blumer’s symbolic interactionism:\n\nWhat helps researchers develop a constructivist grounded theory? How might they shape the data collection and analysis phases? Gaining depth and understanding in their work means that they can fulfill Blumer’s (1969) call for “intimate familiarity” with respondents and their worlds (see also Lofland and Lofland 1984, 1995). In short, constructing constructivism means seeking meanings — both respondents’ meanings and researchers’ meanings.\n\nCharmaz (2000: 524) on the concretization of procedures from what were orginally meant to be guidelines:\n\nGuidelines such as those offered by Strauss and Corbin (1990) structure objectivist grounded theorists’ work. These guidelines are didactic and prescriptive rather than emergent and interactive. Sanders (1995: 92) refers to grounded theory procedures as “more rigorous than thou instructions about how information should be pressed into a mold”. Strauss and Corbin categorize steps in the process with scientific terms such as axial coding and conditional matrix (Strauss 1987; Strauss and Corbin 1990, 1994). As grounded theory methods become more articulated, categorized, and elaborated, they seem to take on a life of their own. Guidelines turn into procedures and are reified into immutable rules, unlike Glaser and Strauss’s (1967) original flexible strategies. By taking grounded theory methods as prescriptive scientific rules, proponents further the positivist cast to obiectivist grounded theory.\n\n\nOn the modes of reasoning behind grounded theory\nKelle (2005) is an overview of the Glaser / Strauss and Corbin split. References to Kelle (2005) have no page numbers since it is published in an online-only journal and does not specify paragraph numbers.\nHighlights a primary impetus behind Glaser and Strauss (1967), which used political analogies to distinguish between “theoretical capitalists” and “proletariat testers”, and unify the field of sociology by de-centering emphasis on theories developed by “great men”.\nA common thread in this paper is sensitivity to the practical challenges of actually doing grounded theory according to Glaser’s approach:\n\nThe infeasibility of an inductivist research strategy which demands an empty head (instead of an “open mind”) cannot only be shown by epistemological arguments, it can also be seen in research practice. Especially novices in qualitative research with the strong desire to adhere to what they see as a basic principle and hallmark of Grounded Theory — the “emergence” of categories from the data — often experience a certain difficulty: in open coding the search for adequate coding categories can become extremely tedious and a subject of sometimes numerous and endless team sessions, especially if one hesitates to explicitly introduce theoretical knowledge. The declared purpose to let codes emerge from the data then leads to an enduring proliferation of the number of coding categories which makes the whole process insurmountable.\n\nKelle (2005) basically takes down the original Glaser and Strauss (1967) and subsequent reflection on theoretecal sensitivity (Glaser 1978). He highlights fundamental contraditions and oversights with regards to the role of theory in grounded theory, specifically with regards to the notion that such research can be accomplished with inductive purity:\n\nConsequently, in the most early version of Grounded Theory the advice to employ theoretical sensitivity to identify theoretical relevant phenomena coexists with the idea that theoretical concepts “emerge” from the data if researchers approach the empirical field with no preconceived theories or hypotheses. Both ideas which have conflicting implications are not integrated with each other in the Discovery book. Furthermore, the concept of theoretical sensitivity is not converted into clear cut methodological rules: it remains unclear how a theoretically sensitive researcher can use previous theoretical knowledge to avoid drowning in the data. If one takes into account the frequent warnings not to force theoretical concepts on the data one gets the impression that a grounded theorist is advised to introduce suitable theoretical concepts ad hoc drawing on implicit theoretical knowledge but should abstain from approaching the empirical data with ex ante formulated hypotheses.\n\nKelle (2005) recognizes that Glaser identified a series of “theoretical families” to help assist with the practical experience of coding. I find it somewhat interesting that many of the terms in these first families are very reminiscient of so-called “natural language”, as used in the wave of cybernets that was contemporary with Glaser (1978) and which largely dealt with “expert systems”.\n\nIn the book “Theoretical Sensitivity” (1978) GLASER presents an extended list of terms which can be used for the purpose of theoretical coding loosely structured in the form of so called theoretical “coding families”. Thereby various theoretical concepts stemming from different (sociological, philosophical or everyday) contexts are lumped together, as for example:\n\nterms, which relate to the degree of an attribute or property (“degree family”), like “limit”, “range”, “extent”, “amount” etc.,\nterms, which refer to the relation between a whole and its elements (“dimension family”), like “element”, “part”, “facet”, “slice”, “sector”, “aspect”, “segment” etc.,\nterms, which refer to cultural phenomena (“cultural family”) like “social norms”, “social values”, “social beliefs” etc.\n\n\nThis is substantiated by other observations by Kelle (2005) that ad hoc coding actually follows implicit theoretical knowledge:\n\nOne of the most crucial differences between GLASER’s and STRAUSS’ approaches of Grounded Theory lies in the fact that STRAUSS and CORBIN propose the utilization of a specified theoretical framework based on a certain understanding of human action, whereas GLASER emphasises that coding as a process of combining “the analyst’s scholarly knowledge and his research knowledge of the substantive field” (1978, p.70) has to be realised ad hoc, which means that it has often to be conducted on the basis of a more or less implicit theoretical background knowledge.\n\nand that the Glaserian approach is better suited for more experienced, rather than novice sociologists, who will have internalized the theory that they then apply in their coding.\nKelle then goes on to address how grounded theory can or can not be applied in alignment with inductivist or hypothetic-deductivist reasoning, and raises abductive reasoning an an alternative means of arriving at legitimate and verifiable conclusions. There is too much detail in the paper to copy here.\nBut here is another nice conclusive gem from the end:\n\nWhereas STRAUSS and CORBIN pay a lot of attention to the question how grounded categories and propositions can be further validated, GLASER’s concept shows at least a gleam of epistemological fundamentalism (or “certism”, LAKATOS 1978) especially in his defence of the inductivism of early Grounded Theory. “Grounded theory looks for what is, not what might be, and therefore needs no test” (GLASER 1992, p.67). Such sentences carry the outmoded idea that empirical research can lead to final certainties and truths and that by using an inductive method the researcher may gain the ability to conceive “facts as they are” making any attempt of further corroboration futile.\n\n\n\nRebuttals by Glaser\nGlaser (2002) constitutes a rebuttal to Kathy Charmaz (2000). As Bryant (2003) points out in his response to Glaser (2002), it is very angry, polemical and irrational. I don’t want to go too in depth with the fundamental problems with Glaser’s response (see Bryant’s paper for the details), but the gist is that Glaser never really got the message about data being inherently constructed by researchers decisions, actions and circumstances. Glaser seems to continue believing in the inherent neutrality of data as a matter of faith.\nThis being said, Glaser (2002) did highlight the large emphasis on descriptive rather than explanatory potential in Charmaz’s approach. This aligns with my own apprehensions when I try to address the relevance of my work. I tend to use the term “articulate” as a way to frame my work as descriptive, but in a way that lends value, and this very fuzzy distinction between the power of identying the shapes and relationships among things and explaining their causes and effects in a generalizable way (i.e., theories, or explanations), still somehow troubles me. I wonder if Glaser is drawing a false distinction here, and through that, a false prioritization of explanation over description as a desired outcome. This would put my mind at ease, as would dismissing Glaser’s dismissal of people who simply don’t know how to do the “real” grounded theory (which, in his mind, include all feminist and critical researchers).\n\n\nOn the functional and pragmatic roots of grounded theory\nI completely agree with this statement from Clarke (2003: 555):\n\nTo address the needs and desires for empirical understandings of the complex and heterogeneous worlds emerging through new world orderings, new methods are requisite (Haraway 1999). I believe some such methods should be epistemologically/ ontologically based in the pragmatist soil that has historically nurtured symbolic interactionism and grounded theory. Through Mead, an interactionist grounded theory has always had the capacity to be distinctly perspectival in ways fully com patible with what are now understood as situated knowledges. This fundamental and always already postmodern edge of a grounded theory founded in symbolic interactionism makes it worth renovating.\n\nThis is super interesting, and really contextualizes how Strauss imagined grounded theory to be useful for him:\n\nSome years ago, Katovich and Reese (1993:400–405) interestingly argued that Strauss’s negotiated order and related work recuperatively pulled the social around the postmodern turn through its methodological [grounded theoretical] recognition of the partial, tenuous, shifting, and unstable nature of the empirical world and its constructedness. I strongly agree and would argue that Strauss also furthered this “postmodernization of the social” through his conceptualizations of social worlds and arenas as modes of understanding the deeply situated yet always also fluid orga nizational elements of negotiations. He foreshadowed what later came to be known as postmodern assumptions: the instability of situations; characteristic changing, porous boundaries of both social worlds and arenas; social worlds seen as mutually constitutive/coproduced through negotiations taking place in arenas; negotiations as central social processes hailing that “things can always be otherwise”; and so on. Significantly, negotiations constitute discourses that also signal micropolitics of power as well as “the usual” meso/macrostructural elements—power in its more fluid forms (e.g., Foucault 1979, 1980). Through integrating the social worlds/arenas/ negotiations framework with grounded theory as a new conceptual infrastructure, I hope to sustain and extend the methodological contribution of grounded theory to understanding and elaborating what has been meant by “the social” in social life — before, during, and after the postmodern turn.\n\nIt also echoes Charmaz’s vision of grounded theory as a powerful too, and Bryant’s (2003) call to “look at what Glaser and Strauss actually did, rather than what they claimed — and continued to claim — they were doing” to uncover “the basis for a powerful research approach”. Bryant (2003) further cites Baszanger and Dodier (1997), who characterize grounded theory as a method “consisting of accumulating a series of individual cases, of analyzing them as a combination between different logics of action that coexist not only in the field under consideration, but even within these individuals or during their encounters”. Bryant (2003) summarizes this by stating that “[t]he aim of such methods is generalization rather than totalization, with the objective of producing”a combinative inventory of possible situations”." }, { - "objectID": "notes/maelstrom-readings.html", - "href": "notes/maelstrom-readings.html", - "title": "Maelstrom reading notes", - "section": "", - "text": "Initial overview of data harmonization procedures, using the Healthy Obesity Project (HOP) as an illustrative case.\nOutlines the technical apparatus, especially for DataShield, but also broadly describes the discursive process of arriving at a DataSchema that is both functional and flexible.\n\nThis description is quite broad and abstracy, seems somewhat ideal and aspirational.\n\nDescribes reliance on international standards, such as the International Labour Organization’s International Standard Classification of Occupations.\n\nIt seems like these are used as black boxes that encapsulate a series of tensions which epidemiologists are unconcerned with; in effect, they simplify the need for stretching the collaborative ties even further than they are already extended, they represent matters out of scope for deeper discursive engagement.\n\nIt is notable that they emphasize that it’s easy to set up and use DataShield and Maelstorm toolkits independently of university IT and that it can be run using RStudio installed on a basic laptop.\n\nMaybe look into the historical context (2013) and the evolving role of university IT in software selection.\n\nThe conclusion states that the HOP project was successful in its harmonization efforts, but does not go as far as to state that it produced meaningful findings as a result of harmonization.\n\nI may take some time to find and read studies that used these data to see what’s what.\nThis seems like the main one: https://doi.org/10.1186/1472-6823-14-9, but these other papers may or not not also be relevant:\n\nhttps://doi.org/10.1016/j.smhl.2021.100263\nhttps://doi.org/10.1007/s10654-014-9977-1\nhttps://doi.org/10.1530/EJE-14-0540\nhttps://doi.org/10.1007/S13679-020-00375-0\nhttps://doi.org/10.1093/eurpub/ckac061" + "objectID": "notes/methodology-notes.html#theoretical-sampling", + "href": "notes/methodology-notes.html#theoretical-sampling", + "title": "Methodology notes", + "section": "Theoretical sampling", + "text": "Theoretical sampling\nSee Kathy Charmaz (2000): 519-520.\nFrom Clarke (2003: 557):\n\nUnique to this approach has been, first, its requiring that analysis begin as soon as there are data. Coding begins immediately, and theorizing based on that coding does as well, however provisionally (Glaser 1978). Second, “sampling” is driven not necessarily (or not only) by attempts to be “representative” of some social body or population (or its heterogeneities) but especially and explicitly by theoretical con cerns that have emerged in the provisional analysis. Such “theoretical sampling” focuses on finding new data sources (persons or things) that can best explicitly ad dress specific theoretically interesting facets of the emergent analysis. Theoretical sampling has been integral to grounded theory from the outset, remains a fundamen tal strength of this analytic approach, and is crucial for the new situational analyses." }, { - "objectID": "notes/maelstrom-readings.html#doiron2013", - "href": "notes/maelstrom-readings.html#doiron2013", - "title": "Maelstrom reading notes", - "section": "", - "text": "Initial overview of data harmonization procedures, using the Healthy Obesity Project (HOP) as an illustrative case.\nOutlines the technical apparatus, especially for DataShield, but also broadly describes the discursive process of arriving at a DataSchema that is both functional and flexible.\n\nThis description is quite broad and abstracy, seems somewhat ideal and aspirational.\n\nDescribes reliance on international standards, such as the International Labour Organization’s International Standard Classification of Occupations.\n\nIt seems like these are used as black boxes that encapsulate a series of tensions which epidemiologists are unconcerned with; in effect, they simplify the need for stretching the collaborative ties even further than they are already extended, they represent matters out of scope for deeper discursive engagement.\n\nIt is notable that they emphasize that it’s easy to set up and use DataShield and Maelstorm toolkits independently of university IT and that it can be run using RStudio installed on a basic laptop.\n\nMaybe look into the historical context (2013) and the evolving role of university IT in software selection.\n\nThe conclusion states that the HOP project was successful in its harmonization efforts, but does not go as far as to state that it produced meaningful findings as a result of harmonization.\n\nI may take some time to find and read studies that used these data to see what’s what.\nThis seems like the main one: https://doi.org/10.1186/1472-6823-14-9, but these other papers may or not not also be relevant:\n\nhttps://doi.org/10.1016/j.smhl.2021.100263\nhttps://doi.org/10.1007/s10654-014-9977-1\nhttps://doi.org/10.1530/EJE-14-0540\nhttps://doi.org/10.1007/S13679-020-00375-0\nhttps://doi.org/10.1093/eurpub/ckac061" + "objectID": "notes/methodology-notes.html#interviews", + "href": "notes/methodology-notes.html#interviews", + "title": "Methodology notes", + "section": "Interviews", + "text": "Interviews\nSee\nFrom Charmaz (2000: 525):\n\nA constructivist approach necessitates a relationship with respondents in which they can cast their stories in their terms. It means listening to their stories with openness to feeling and experience. … Furthermore, one-shot interviewing lends itself to a partial, sanitized view of experience, cleaned up for public discourse. The very structure of an interview may preclude private thoughts and feelings from emerging. Such a structure reinforces whatever proclivities a respondent has to tell only the public version of the story. Researchers’ sustained involvement with research participants lessens these problems.\n\nFontana and Frey (2000) spend some time writing about the emergence of an “interview society”, whereby interviews are commonly used to seek various forms of biographical information. They cite Gubrium and Holstein (1998), who noted that “the interview has become a means of contemporary storytelling, where persons divulge life accounts in response to interview inquiries”. They then go over a brief history of interviewing in the context of sociological research, which largely tracks the values underlying positivist and postmodernist transitions as you might expect.\nYin (2014: 110-113) differentiates between three kinds of interviews:\n\nProlonged interviews: Usually over two hours long, or over an extended period of time covering multiple sittings.\n\nYou can ask interviewees about their interpretations and opinions about people and events or their insights, explanations, and meanings related to certain occurrences. You can then use such propositions as the basis for further inquiry, and the interviewee can suggest other persons for you to interview, as well as other sources of evidence. The more that an interviewee assists in this manner, the more that the role may be considered one of an “informant” rather than a participant. Key informants are often critical to the success of a case study. Such persons can provide you with insights into a matter and also give you access to other interviewees who may have corroboratory or contrary evidence.\n\nShorter interviews: More focused, around one hour, open-ended and conversational but largely following the protocol.\n\nA major purpose of such an interview might simply be to corroborate certain findings that you already think have been established, but not to ask about other topics of a broader, open-ended nature. In this situation, the specific questions must be carefully worded, so that you appear genuinely naive about the topic and allow the interviewee to provide a fresh commentary about it; in contrast, if you ask leading questions, the corroboratory purpose of the interview will not have been served. … As an entirely different kind of example, your case study protocol might have called for you to pay specific attention to an interviewee’s personal rendition of an event. In this case, the interviewee’s perceptions and own sense of meaning are the material to be understood. … In both [situations], you need to minimize a methodological threat created by the conversational nature of the interview.\n\nSurvey interviews: A more structured questionnaire. Usually works best as one component of multiple sources of evidence.\n\n\nStructured interviewing\nFrom Fontana and Frey (2000: 649-651):\nInterviewers ask respondents a series of preestablished questions with a limited set of response categories. The interview records responses according to a preestablished coding scheme.\nInstructions to interviewers often follow these guidelines:\n\nNever get involved in long explanations of the study; use the standard explanation provided by the supervisor.\nNever deviate from the study introduction, sequence of questions, or question wording.\nNever let another person interrupt the interview; do not let another person answer for the respondent or offer his or her opinions on the question.\nNever suggest an answer or agree or disagree with an answer. Do not give the respondent any idea of your personal views on the topic of the question or the survey.\nNever interpret the meaning of a question; just repeat the question and give instructions or clarifications that are provided in training or by the supervisors.\nNever improvise, such as by assing answer categories or making wording changes.\n\nThe interviewer must establish a “balanced rapport”, being casual and friendly while also directive and impersonal. Interviewers must also perfect a style of “interested listening” that rewards respondents’ participation but does not evaluate their responses.\nFrom Fontana and Frey (2000: 651):\n\nThis kind of interview often elicits rational responses, but it overlooks or inadequately assesses the emotional dimension.\n\n\n\nGroup interviews\nFrom Fontana and Frey (2000: 651-652):\nCan be used to test a methodological technique, try out a definition of a research problem or to identify key informants. Pre-testing a questionnaire or survey design.\nCan be used to aid respondents’ recall of specific events or to stimulate embellished descriptions of events, or experiences shared by members of a group.\nIn formal group interviews, participants share views through the coordinator.\nLess formal group interviews are meant to establish the widest range of meaning and interpretation on a topic, and the objective is “to tap intersubjective meaning with depth and diversity”.\n\n\nUnstructured interviewing\nFrom Fontana and Frey (2000: 652-657):\nThe essence of an unstructured interview is establishing a human-to-human relation with the respondent and a desire to understand rather than to explain.\nFontana and Frey (2000) then goes on with some practical guidance on how to engage in unstructured interviews, largely concerned with how to access a community and relate with respondents.\n\n\nPostmodern takes on interviewing\nFontana and Frey (2000) address some “new” takes on interviewing emerging from the postmodern turn. I kinda think there is some potential behind approaches that emphasize interviews as negotiated accomplishment, or product of communal sensemaking between interviewer and respondent. I think it could be really helpful in the context of my research, which is very concerned with drawing out tensions that respondents have in mind but are not really able to articulate in a systematic way.\nHowever, Fontana and Frey (2000) also draws attention to crticism of highly engaged interviewing approaches, which seem to equate closeness with the respondent as getting closer to their “true self”, and which may not actually be fixed (especially in the context of the artificial environment of the interview setting). Critiques of such “closeness” use the term “romantic” or “crusading” as epithets. Moreover, there is additional reference to the culturally-embedded assumption that interviews are necessarily valuable sources of information, as if speaking ones mind can adequately convey one’s thoughts and experiences — this is criticized as a particularly western approach to information extraction surrounding internalized and externalized thoughts and behaviour, as instilled through participation in the “interview society” addressed earlier in the text.\n\n\nTranscribing\nThis section describes how I transcibe interviews and accounts for the decisions to encode certain things and not others. It goes on to explains the procedures for transcribing spoken dialog into textual formats, including the notation applied to encode idiosyncratic elements of conversational speech.\n\nTranscript notation\nDerived from the transcription protocol applied for the E-CURATORS project.\n\n\nCleaning audio\nTo clean the audio:\n\nSelect a clip that is representative of a single source of background noise, and then filter that wavelength throughout the entire audio file.\nAfter selecting the clip, go to Effect >> Noise Reduction and select Get Noise Profile, then press OK.\nClose the noise reduction menu, select the entire range of audio using the keyboard shortcut Command + A.\nThen go back to the noise reduction window (Effect >> Noise Reduction) to apply the filter based on the noise profile identified for the noisy clip.\nExport the modified audio file to the working directory (File >> Export >> Export as .WAV).\nUse ffmpeg to replace the dirty audio track with the clean one:\n\n ffmpeg -i dirty.mp4 -i clean.wav -c:v copy -map 0:v:0 -map 1:a:0 clean.mp4" }, { - "objectID": "notes/maelstrom-readings.html#doiron2017", - "href": "notes/maelstrom-readings.html#doiron2017", - "title": "Maelstrom reading notes", - "section": "Doiron et al. (2017)", - "text": "Doiron et al. (2017)\n\nAn overview of the key software that facilitates data harmonization practices under Maelstrom, also briefly touched upon in Doiron et al. (2013).\nPage 1373 refers to graphical and programmatic interfaces and assumes certain roles and tasks associated with each.\nBriefly describes its use by the Canadian Longitudinal Study on Aging (CLSA), the Canadian Partnership for Tomorrow Project (CPTP) and InterConnect, primarily by describing the range and quantity of data that these systems manage in each case.\n\n\nOpal provides a centralized web-based data management system allowing study coordinators and data managers to securely import/export a variety of data types (e.g. text, nu merical, geolocation, images, videos) and formats (e.g. SPSS, CSV) using a point-and-click interface. Opal then converts, stores and displays these data under a standar dized model.\n\n\nMica is used to create websites and metadata portals for individual epidemiological studies or multi-study consor tia, with a specific focus on supporting observational co hort studies. The Mica application helps data custodians and study or network coordinators to efficiently organize and disseminate information about their studies and net works without significant technical effort." + "objectID": "notes/methodology-notes.html#field-notes", + "href": "notes/methodology-notes.html#field-notes", + "title": "Methodology notes", + "section": "Field notes", + "text": "Field notes\nSee (Yin 2014: 124-125)" }, { - "objectID": "notes/maelstrom-readings.html#fortier2010", - "href": "notes/maelstrom-readings.html#fortier2010", - "title": "Maelstrom reading notes", - "section": "Fortier et al. (2010)", - "text": "Fortier et al. (2010)\n\nA very grandiose paper presenting the grand vision for DataSHaPER, which would eventually become Maelstrom.\n\nLots of co-authors!\n\nInvokes the pan-European EPIC project (European Prospective Investigation into Cancer and Nutrition), which faced numerous data synthesis challenges despite its proactive effort to coordinate work across numerous research centres.\n\n\nTwo complementary approaches may be adopted to support effective data synthesis. The first one principally targets ‘what’ is to be synthesized, whereas the other one focuses on ‘how’ to collect the required information. Thus: (i) core sets of information may be identified to serve as the foundation for a flexible approach to harmonization; or (ii) standard collection devices (questionnaires and stand ard operating procedures) may be suggested as a required basis for collection of information.\n\n\nDataSHaPER is an acronym for DataSchema and Harmonization Platform for Epidemiological Research.\n\n\nIn an ideal world, information would be ‘prospectively harmonized’: emerging studies would make use, where possible, of harmonized questionnaires and standard operating procedures. This enhances the potential for future pooling but entails significant challenges —- ahead of time -— in developing and agree ing to common assessment protocols. However, at the same time, it is important to increase the utility of existing studies by ‘retrospectively harmonizing’ data that have already been collected, to optimize the subset of information that may legitimately be pooled. Here, the quantity and quality of infor mation that can be pooled is limited by the heterogeneity intrinsic to the pre-existing differences in study design and conduct.\n\nCompares prospective and retrospective harmonizatiom, with the former being presented as ideal, and the latter being a pragmatic reconciliation in acknowledgement that the former is essentially impossible to achieve.\n\nDataSHaPER is strikingly similar to OCHRE:\n\nXML-based data structures\nGenesis of a generic and ultimately optional base-level schema that illustrates the kind of data that the data structure may hold in ways that are immediately recognizable to all practitioners (at OCHRE it was associations between contexts and finds)\nSeparate harmonization platform where users can edit and manipulate records and associations between them\n\n\n\nThe question ‘What would constitute the ultimate proof of success or failure of the DataSHaPER approach’ needs to be addressed. Such proof will necessarily accumulate over time, and will involve two fundamental elements: (i) ratification of the basic DataSHaPER approach; and (ii) confirmation of the quality of each individual DataSHaPER as they are developed and/or extended. An important indication of the former would be provided by the widespread use of our tools. However, the ultimate proof of principle will necessarily be based on the generation of replicable scientific findings by researchers using the approach. But, for such evidence to accumulate it will be essential to assure the quality of each individual DataSHaPER. Even if the fundamental approach is sound, its success will depend critically on how individual DataSHaPERs are constructed and used. It seems likely that if consistency and quality are to be assured in the global development of the approach, it will be necessary for new DataSHaPERs to be formally endorsed by a central advisory team." + "objectID": "notes/methodology-notes.html#focus-groups", + "href": "notes/methodology-notes.html#focus-groups", + "title": "Methodology notes", + "section": "Focus groups", + "text": "Focus groups\nFrom Kitzinger (1994); Wilkinson (1998); Parker and Tritter (2006); Morgan and Hoffman (2018).\n\nCrucially, focus groups involve the interaction of group participants with each other as well as with the moderator, and it is the collection of this kind of interactive data which distinguishes the focus group from the one-to-one interview (c.f. Morgan 1988), as well as from procedures which use multiple participants but do not permit interactive discussion (c.f. Stewart and Shamdasani 1990). The ‘hallmark’ of focus groups, then, is the ‘explicit use of group interaction to produce data and insights that would be less accessible without the interaction found in a group’ (Morgan 1997: 2).\n\nWilkinson (1998) identifies three key features of focus group methods:\n\nproviding access to participants’ own language, concepts and concerns;\n\n\n“The relatively free flow of discussion and debate between members of a focus group offers an excellent opportunity for hearing ‘the language and vernacular used by respondents’”\nUseful for gaining insight into participants’ conceptual worlds on their own terms\n“Sometimes, the participants even offer the researcher a ‘translation’ of unfamiliar terms or concepts”\n“Focus group interactions reveal not only shared ways of talking, but also shared experiences, and shared ways of making sense of these experiences. The researcher is offered an insight into the commonly held assumptions, concepts and meanings that constitute and inform participants’ talk about their experiences.”\n\n\nencouraging the production of more fully articulated accounts;\n\n\n“In focus groups people typically disclose personal details, reveal discrediting information, express strong views and opinions. They elaborate their views in response to encouragement, or defend them in the face of challenge from other group members: focus groups which ‘take off’ may even, like those run by Robin Jarrett (1993: 194) have ‘the feel of rap sessions with friends’.”\n“Even when bcus group participants are not acquainted in advance, the interactive nature of the group means that participants ask questions of, disagree with, and challenge each other, thus serving ‘to elicit the elaboration of responses’ (Merton 1987: 555).”\n“Other ethical issues in focus group research stem from group dynamics, insofar as participants can collaborate or collude effectively to intimidate and/or silence a particular member, or to create a silence around a particular topic or issue, for example. In such cases, it falls to the group moderator to decide whether/how to intervene, and it can be difficult to balance such conflicting goals as ensuring the articulation of accounts, supporting individuals, and challenging offensive statements.”\n\n\nand offering an opportunity to observe the process of collective sense-making.\n\n\nFocus groups also offer an opportunity for researchers to see exactly how views are constructed, expressed, defended and (sometimes) modified during the course of conversations with others, 1.e. to observe the process of collective sense-making in action. This is likely to be of particular interest to researchers working within a social constructionist framework, who do not view beliefs, ideas, opinions and understandings as generated by individuals in splendid isolation, but rather as built in interaction with others, in specific social contexts: as Radley and Billig (1996: 223) say, ‘thinking is a socially shared activity’. In a focus group, people are confronted with the need to make collective sense of their individual experiences and beliefs (Morgan and Spanish 1984: 259).\n\nWilkinson (1998) addresses opportunities and challenges in analyzing focus group data using content analysis and ethnographic techniques.\nWith regards to content analysis:\n\nthe main advantages of content analysis are to allow for a relatively systematic treatment of the data and to enable its presentation in summary form. … the researcher has first to decide on the unit of analysis: this could be the whole group, the group dynamics, the individual participants, or the participants’ utterances\n\n\nMorgan (1997) proposes three distinct ways of coding focus group data: noting whether each group discussion contains a given code; noting whether each participant mentions a given code; and noting all mentions of a given code (i.e. across groups or participants). Once the data have been coded in one (or more) of these ways, the question of whether to quantify them is a further issue. Morgan (1993) argues the value of simple ‘descriptive counts’ of codes (stopping short of using inferential statistical tests, whose assumptions are unlikely to be met in focus groups)\n\nAnd with regards to ethnographic analysis:\n\nIts main advantage is to permit a detailed interpretative account of the everyday social processes of communication, talk and action occurring within the focus group. The key issue in ethnographic analysis is how to select the material to present (whether this is framed up as ‘themes’, ‘discourses’, or simply as illustrative quotations), without violating the ‘spirit’ of the group, and without losing sight of the specific context within which the material was generated. … A particular challenge is how to preserve the interactive nature of focus group data: a surprising limitation of published focus group research is the rarity with which group interactions are analysed and reported ( c.f. Carey and Smith 1994, Kitzinger 1994a). Extracts from focus group data are most commonly presented as if they were one-to-one interview data, often with no indication that more than one person is present; still more rarely does interaction per se constitute the analytic focus.\n\n\nKitzinger (1994) elaborates on the interactive nature of focus groups, which she identifies as the methods’ core feature.\n\nFocus groups are group discussions organised to explore a specific set of issues sucb as people’s views and experiences … The group is ‘focused’ in the sense that it involves some kind of collective activity - such as viewing a film, examining a single health education message or simply debating a particular set of questions.\n\n\nEven when group work is explicitly included as part of the research it is often simply employed as a convenient way to illustrate a theory generated by other methods or as a cost-effective technique for interviewing several people at once. Reading some such reports it is hard to believe that there was ever more than one person in the room at the same time. This criticism even applies to many studies which explicitly identify their methodology as ‘focus group discussion’ — in spite of the fact that the distinguishing feature of focus groups is supposed to be the use of interaction as part of the research data.\n\n\nIt would be naive, however, to assume that group data is by definition ‘natural’ in the sense that it would have occurred without the group having been convened for this purpose. It is important to note that although, at times, the focus groups may approximate to participant observation the focus groups are artifidally set up situations. Rather than assuming the group session unproblematically and inevitably reflects ‘everyday interactions’ (although sometimes it will) the group sbould be used to encourage people to engage with one another, verbally formulate their ideas and draw out the cognitive structures which previously have been unarticulated.\n\nOn the active role of the facilitator:\n\nSessions were conducted in a relaxed fashion with minimal intervention from the facilitator - at least at first. This allowed the facilitator to ‘find her feet’ and permitted the research participants to set the priorities. However, the researcber was never passive. Trying to maximise interaction between participants could lead to a more interventionist style: urging debate to continue beyond the stage it might otberwise have ended, challenging people’s taken for granted reality and encouraging them to discuss the inconsistencies both between participants and within their own thinking.\n\nOn the role of the presentation or activity:\n\nSuch exercises not only provided invaluable data from each group but allow for some cross-compadsons between groups. Each discussion session has its own dynamic and direction — when it comes to analysis it is extremely useful to have a common external reference point such as that provided by the card game or the use of vignettes (Kban and Manderson 1992).\n\nOn complementary interactions:\n\nThe excbange between the research participants not only allows the researcher to understand which advertisement they are talking about but to gather data on their shared perception of that image.\n\n\nBrainstorming and loose word association was a frequent feature of the research sessions.\n\n\npeople’s knowledge and attitudes are not entirely encapsulated in reasoned responses to direct questions. Everyday forms of communication such as anecdotes, jokes or loose word association may tell us as much, if not more, about what people ‘know’. In this sense focus groups ‘reach the parts that other methods cannot reach’ - revealing dimensions of understanding that often remain untapped by the more conventional one-to-one interview or questionnaire.\n\n\nIn addition to the advantages discussed above fo exampie, discussing whether they had the ‘right’ to know if another child in the piay group bad had virus asserted that ‘you think of your own first’. It was this phrase, and these sort of sentiments, which seemed to capture their consent and resulted in nods of agreement round the group and assertions that ‘that’s right’ and ‘of course’. Indeed, it was often the strength of the collective reaction that highlighted the specific context within which the research participants experienced AIDS information.\n\nOn argumentative interactions:\n\nthe group process however, is not only about consensus and the articulation of group norms and experiences. Differences between individuals within the group are equally important and, in any case, rarely disappear from view. Regardless of how they are selected, the research participants in any one group are never entirely homogenous. Participants do not just agree with each other — they also misunderstand one another, question one another, try to persuade each other of the justice of their own point of view and sometimes they vehemently disagree.\n\n\nSuch unexpected dissent led them to clarify why they thpught as they did, often identifying aspects of their personal experience which bad altered their opinions or specific occasions which had made them re-think their point of view. Had the data been collected by interviews the researcber might have been faced with ‘arm chair’ theorizing about the causes of such difference but in a focus group these can be explored ‘in situ’ with the help of the research participants.\n\n\nClose attention to the ways in which research participants tell stories to one another aiso prevents the researcber from assuming that she knows ‘the meaning’ of any particular anecdote or account.\n\n\nParker and Tritter (2006) discuss “key issues relating to the complexity and necessity of considering sampling issues within the context of focus group research and the implications this has for the collection and analysis of resultant data.”\nIdentifies some common logistical impetus for focus groups, relating to acquisition of funding:\n\nIncreasing pressure from research funding organizations to adopt multiple-method research strategies and the fact that focus groups generate far more data than a range of other methods in relation to face-to-face contact between researchers and participants, has added to [the method’s popularity].\n\nCalls out the conflation between focus groups and group interviews:\n\nA similarly pervasive trend in and around academic discussion of qualitative research methods is that focus groups are sometimes seen as synonymous with group interviews and it is this issue which constitutes our second point of contention. … In keeping with the views of a number of other writers in this field, we are of the opinion that there is a fundamental difference between these two research techniques and that the critical point of distinction surrounds the role of the researcher and her/ his relationship to the researched (Smithson, 2000). In group interviews the researcher adopts an ‘investigative’ role: asking questions, controlling the dynamics of group discussion, often engaging in dialogue with specific participants. This is premised on the mechanics of one-to-one, qualitative, in-depth interviews being replicated on a broader (collective) scale. A relatively straightforward scenario ensues: the researcher asks questions, the respondents relay their ‘answers’ back to the researcher. In focus groups the dynamics are different. Here, the researcher plays the role of ‘facilitator’ or ‘moderator’; that is, facilitator/moderator of group discussion between participants, not between her/himself and the participants. Hence, where focus groups are concerned, the researcher takes a peripheral, rather than a centre-stage role for the simple reason that it is the inter-relational dynamics of the participants that are important, not the relationship between researcher and researched (see Kitzinger, 1994a; Johnson, 1996).\n\n\nPart of the problem of achieving this kind of interactional synergy in data collection is that, despite their collective interests, participants may not always be keen to engage with each other, or alternatively, may know each other so well that interaction is based on patterns of social relations that have little to do with the research intent of the focus group. The need to consider the impact on interaction of the constitution of the focus group requires that close attention be paid to methods (and outcomes) of recruit ment. … issues of sampling and selection are likely to prove crucial in relation to the form and quality of interaction in a focus group and therefore the kinds of data one gathers and the extent to which participants share their opinions, attitudes and life experiences.\n\nA large part of what follows is an extremely honest account of the challenges involved in recruiting participants to sit for focus groups. The pressure on students to recruit their friends impacted their views on what views were represented and not represented. Moreover, they draw attention to the role of timing and convenience in recruitment strategy, and representativeness of expressed viewpoints.\nI found it hard to take notes on this since it comes off as a continuous stream of conciousness. However it should still be read from top to bottom to get a sense of practical challenges that are not commonly addressed, save for among veterans and in private circles.\n\nMorgan and Hoffman (2018)" }, { - "objectID": "notes/maelstrom-readings.html#fortier2011", - "href": "notes/maelstrom-readings.html#fortier2011", - "title": "Maelstrom reading notes", - "section": "Fortier et al. (2011)", - "text": "Fortier et al. (2011)\nThis paper responds to Hamilton et al. (2011), which presents an effort to devise a standardized nomenclature. The response is basically to advocate for a more flexible approach, rather than a stringent one promoted by Hamilton et al. (2011). It draws extensively from concepts published in the foundational paper by Fortier et al. (2010).\n\nTwo complementary approaches to harmonization may be adopted to support effective data synthesis or comparison across studies. The first approach makes use of identical data collection tools and procedures as a basis for harmoni zation and synthesis. Here we refer to this as the ‘‘stringent’’ approach to harmonization. The second approach is con sidered ‘‘flexible’’ harmonization. Critically, the second ap proach does not demand the use of identical data collection tools and procedures for harmonization and synthesis. Rather, it has to be based on sound methodology to ensure inferential equivalence of the information to be harmonized. Here, standardization is considered equivalent to stringent harmonization. It should, however, be noted that the term standard is occasionally employed to refer to common con cepts or comparable classification schemes but does not necessarily involve the use of identical data collection tools and procedures (12, 13).\n\nThis directly parallels the distinction made in Fortier et al. (2010) between “ideal” prospective and more pragmatic retrospective approaches to data harmonization.\n\nSynthesis of data using a flexible harmonization approach may be either prospective or retrospective. To achieve flexible prospective harmonization, investigators from several studies will agree on a core set of variables (or measures), compatible sets of data collection tools, and standard operating procedures but will allow a certain level of flexibilit in the specific tools and procedures used in each study (16, 17). Retrospective harmonization targets synthesis of information already collected by existing legacy studies (15, 18, 19). As an illustrative example, using retrospective harmonization, researchers will define a core set of variables (e.g., body mass index, global level of physical activity) and, making use of formal pairing rules, assess the potential for each participating study to create each variable (15). The ability to retrospectively harmonize data from existing studies facilitates the rapid generation of new scientifi knowledge.\n\nI wonder why there is no example provided for prospective data harmonization. Is it because it is ideal and not realistic? I’d argue that it is simply what occurs within individual projects." + "objectID": "notes/methodology-notes.html#coding", + "href": "notes/methodology-notes.html#coding", + "title": "Methodology notes", + "section": "Coding", + "text": "Coding\nThese notes are largely derived from my reading of Saldaña (2016), provides a practical overview of what coding entails and specific methods and techniques.\nCoding as component of knowledge construction:\n\nCoding is an intermediate step, “the”critical link” between data collection and their explanation or meaning” (from Kathy Charmaz 2001; as quoted in Saldaña 2016: 4).\n“coding is usually a mixture of data [summation] and data complication … breaking the data apart in analytically relevant ways in order to ead toward further questions about the data” (from Coffey and Atkinson 1996: 29-31; as quoted and edited in Saldaña 2016: 9).\n\nThis relates to the paired notions of decodng when we reflect on a passage to decipher its core meaning, and encoding when we determine its appropriate code and label it (Saldaña 2016: 5).\n\nCoding “generates the bones of your analysis. … [I]ntegration will assemble those bones into a working skeleton” (from Kathy Charmaz 2014: 113; as quoted in Saldaña 2016: 9).\nTo codify is to arrange things in a systematic order, to make something part of a system or classification, to categorize\n\nWhat I sometimes refer to as arranging the code tree\nWhat Saldaña (2016) refers to as categories, I tend to refer to as stubs\n\nCategories are arranged into themes or concepts, which in turn lead to assertions or theories\n\n\nPre-coding techniques\nSaldaña (2016) identifies several techniques for formatting the data to make them easier to code, but also to imbue meaning in the text.\n\nData layout\n\nSeparation between lines or paragraphs may hold significant meaning\nPutting interviewer words in square brackets or capital letters\n\nSemantic markup\n\nBold, italics, underline, highlight\nMeant to identify “codable moments” worthy of attention (Boyatzis 1998; as referenced in Saldaña 2016: 20)\nRelates to Saldaña’s (2016: 22) prompt: “what strikes you?”\n\nPreliminary jottings\n\nTri-column exercise with the text on the left, first impression or preliminary code in the middle, and code on the right, after Liamputtong and Ezzy (2005): 270-273.\n\nAsking questions back to the interviewer, or participating in an imagined dialogue\n\nI imagine this might be useful in situations where the time to hold an interview is quite limited and I have to work with limited responses that don’t touch on everything I want to cover\nThe form of questions maintains my tentativity, my unwillingness to commit or assume their responses, and opens the door for their own responses in rebuttal\n\n\nFollowing Emerson et al. (2011: 177), Saldaña (2016: 22) identifies a few key questions to keep in mind while coding:\n\nWhat are people doing? What are they trying to accomplish?\nHow, exactly, do they do this? What specific means and/or strategies do they use?\nHow do members talk about, characterize, and understand what is going on?\nWhat assumptions are they making?\nWhat do I see going on here?\nWhat did I learn from these notes?\nWhy did I include them?\nHow is what is going on here similar to, or different from, other incidents or events recorded elsewhere in the fieldnotes?\nWhat is the broader import or significance of this incident or event? What is it a case of?\n\n\n\nCoding frameworks\nFrom Clarke (2003: 558) on process coding:\n\nIn a traditional grounded theory study, the key or basic social process is typically articulated in gerund form connoting ongoing action at an abstract level. Around this basic process are then constellated the particular and distinctive conditions, strategies, actions, and practices engaged in by human and nonhuman actors in volved with/in the process and their consequences. For example, subprocesses of disciplining the scientific study of reproduction include formalizing a scientific disci pline, gleaning fiscal support for research, producing contraceptives and other techno scientific products, and handling any social controversies the science provokes (such as cloning and stem cell research)." }, { - "objectID": "notes/maelstrom-readings.html#fortier2017", - "href": "notes/maelstrom-readings.html#fortier2017", - "title": "Maelstrom reading notes", - "section": "Fortier et al. (2017)", - "text": "Fortier et al. (2017)\nExplicit statement regarding the rationale and presumed benefits of harmonization right in the first paragraph:\n\nThe rationales underpinning such an approach include ensuring: sufficient statistical power; more refined subgroup analysis; increased exposure hetero geneity; enhanced generalizability and a capacity to under take comparison, cross validation or replication across datasets. Integrative agendas also help maximizing the use of available data resources and increase cost-efficiency of research programmes.\n\nSummarized in bullet points:\n\nensuring sufficient statistical power\nmore refined subgroup analysis\nincreased exposure heterogeneity\nenhanced generalizability\na capacity to undertake comparison, cross validation or replication across datasets.\nmaximizing the use of available data resources\nincrease cost-efficiency of research programmes\n\nClearly defines harmonization and its benefits:\n\nEssentially, data harmonization achieves or improves comparability (inferential equivalence) of similar measures collected by separate studies.\n\nAdds an additional argument for retrospective harmonization on top of prior discussion of retrospective/prospective approaches (cf. Fortier et al. (2010); Fortier et al. (2011)):\n\nRepeating identical protocols is not necessarily viewed as providing evidence as strong as that obtained by exploring the same topic but using different designs and measures.\n\nAlso relates retrospective harmonization from systematic meta reviews. In fact, the paper basically responds to calls for more structured guidelines for data harmonization, similar to those that had been produced to support structured metareviews in the years prior to this publication. The authors identify several papers that have done similar guidelines or reports on harmonization practices, which they claim are too broad. Those papers include:\n\nRolland et al. (2015)\n\nhttps://doi.org/10.1093/aje/kwv133\n\nSchaap et al. (2011)\n\nhttps://doi.org/10.1186/1471-2474-12-272\n\nBennett et al. (2011)\n\nhttps://doi.org/10.1002/gepi.20564\n\nHohmann et al. (2012)\n\nThis paper reports the findings of a systematic inquiry made to data harmonization initiatives, whose data comprise responses to a questionnaire. The findings indicate that procedures were more attentively follows during earlier stages, such as when matching and aligning available data with the project’s designated scope. However, procedures were less sound with regards to documenting procedures, validating the results of data processing, and dissemination strategy. There is a notable division between work that occurs before and after people actually begin handling the data, which indicates a tension between aspirational plans and tangible experiences.\n\nRespondents were asked to delineate the specific procedures or steps undertaken to generate the harmonized data requested. Sound procedures were generally described; however, the terminologies, sequence and technical and methodological approaches to these procedures varied considerably. Most of the procedures mentioned were related to defining the research questions, identifying and selecting the participating studies (generally not through a systematic approach), identifying the targeted variables to be generated and processing data into the harmonized variables. These procedures were reported by at least 75% of the respondents. On the other hand, few reported steps related to validation of the harmonized data (N=4; 11.8%), documentation of the harmonization process (N=5; 14.7%) and dissemination of the harmonized data outputs (N=2; 5.9%).\n\nThe paper summarizes some specific “potential pitfalls” reported by respondents to their survey:\n\nensuring timely access to data;\nhandling dissimilar restrictions and procedures related to individual participant data access;\nmanaging diversity across the rules for authorship and recognition of input from study-specific investigators;\nmobilizing sufficient time and resources to conduct the harmonization project;\ngathering information and guidance on harmonization approaches, resources and techniques;\nobtaining comprehensive and coherent information on study-specific designs, standard operating procedures, data collection devices, data format and data content;\nunderstanding content and quality of study-specific data;\ndefining the realistic, but scientifically acceptable, level of heterogeneity (or content equivalence) to be obtained;\ngenerating effective study-specific and harmonized datasets, infrastructures and computing capacities;\nprocessing data under a harmonized format taking into account diversity of: study designs and content, study population, synchronicity of measures (events measured at different point in time or at different intervals when repeated) etc;\nensuring proper documentation of the process and decisions undertaken throughout harmonization to ensure transparency and reproducibility of the harmonized datasets;\nmaintaining long-term capacities supporting dissemination of the harmonized datasets to users.\n\nIt’s not made clear how these responses were distributed among respondents.\nThe authors then identify several absolute essential requirements needed to achieve success:\n\nCollaborative framework: a collaborative environment needs to be implemented to ensure the success of any harmonization project. Investigators involved should be open to sharing information and knowledge, and investing time and resources to ensure the successful implementation of a data-sharing infrastructure and achievement of the harmonization process.\nExpert input: adequate input and oversight by experts should be ensured. Expertise is often necessary in: the scientific domain of interest (to ensure harmonized variables permit addressing the scientific question with minimal bias); data harmonization methods (to support achievement of the harmonization procedures); and ethics and law (to address data access and integration issues).\nValid data input: study-specific data should only be harmonized and integrated if the original data items collected by each study are of acceptable quality.\nValid data output: transparency and rigour should be maintained throughout the harmonization process to ensure validity and reproducibility of the harmonization results and to guarantee quality of data output. The common variables generated necessarily need to be of acceptable quality.\nRigorous documentation: publication of results generated making use of harmonized data must provide the information required to estimate the quality of the process and presence of potential bias. This includes a description of the: criteria used to select studies; process achieved to select and define variables to be harmonized; procedures used to process data; and characteristics of the study-specific and harmonized dataset(s) (e.g. attribute of the populations).\nRespect for stakeholders: all study-specific as well as network-specific ethical and legal components need to be respected. This includes respect of the rights, intellectual property interests and integrity of study participants, investigators and stakeholders.\n\nThe authors describe how they arrived at guidelines following the results of this study:\n\nA consensus approach was used to assemble information about pitfalls faced during the harmonization process, establish guiding principles and develop the guidelines. The iterative process (informed by workshops and case studies) permitted to refine and formalize the guide lines. The only substantive structural change to the initial version proposed was the addition of specific steps relating to the validation, and dissemination and archiving of harmonized outputs. These steps were felt essential to em phasize the critical nature of these particular issues.\n\nThe paper outlines a checklist of stages that data harmonization initiatives need to go through to produce ideal outcomes. For each task, they describe a scenario in which the task can be said to be complete, whhich resembles an ideal outcome. This is described in the paper, summarized in a table, and more comprehensively documented in the supplementary materials.\nAlso worth noting, this paper includes a list of harmonization initiatives that I may consult when selecting cases. I’m not quite sure how useful it will be since the findings don’t really break down the distribution of responses in any detail, but maybe the authors have done this analysis and not published it." + "objectID": "notes/methodology-notes.html#memos", + "href": "notes/methodology-notes.html#memos", + "title": "Methodology notes", + "section": "Memos", + "text": "Memos\nCharmaz (2014: 162) dedicates Chapter 7 to memo-writing, which she frames as “the pivotal intermediate step between data collection and writing drafts of papers.” She locates the power of memo-writing as the prompt to analyze the data and codes early in the research process, which requires the researcher to pause and reflect.\nCharmaz (2014: 162):\n\nMemos catch your thoughts, capture the comparisons and connections you make, and crystallize questions and directions for you to pursue.\n\nCharmaz (2014: 162):\n\nMemo-writing creates an interactive space for conversing with yourself about your data, codes, ideas, and hunches. Questions arise. New ideas occur to you during the act of writing. Your standpoints and assumptions can become visible. You will make discoveries about your data, emerging categories, the developing frame of your analysis — and perhaps about yourself.\n\nCharmaz (2014: 164):\n\nMemo-writing encourages you to stop, focus, take your codes and data apart, compare them, and define links between them. Stop and catch meanings and actions. Get them down on paper and into your computer files.\n\nCharmaz (2014: 164):\n\nMemos are your path to theory constriction. They chronicle what you grappled with and learned along the way.\n\nCharmaz (2014: 165-168) distinguishes between memo-writing and journaling. The former is meant to be more analytical, whereas the latter is more of an account of a direct experience, including significant memories or recollections of moments that stood out (and reflection on why they stood out).\nCharmaz (2014: 171) indicates that “[n]o single mechanical procedure defines a useful memo. Do what is possible with the material you have.” She then lists a few possible approaches to memo-writing:\n\nDefine each code or category by its analytic properties\nSpell out and detail processes subsumed by the codes or categories\nMake comparisons between data and data, data and codes, codes and codes, codes and categories, categories and categories\nBring raw data into the memo\nProvide sufficient empirical evidence to support your definitions of the category and analytic claims about it\nOffer conjectures to check in the field setting(s)\nSort and order codes and categories\nIdentify gaps in the analysis\nInterrogate a code or category by asking questions of it.\n\nCharmaz (2014: 171) draws special attention on bringing data into the memo as a way to more concretely “ground” the abstract analysis in the data and lay the foundation for making claims about them:\n\nIncluding verbatim material from different sources permits you to make precise comparisons right in the memo. These comparisons enable you to define patterns in the empirical world. Thus, memo-writing moves your work beyond individual cases.\n\nThrough a detailed example over the prior several pages, Charmaz (2014: 178) reflects on how memos may “[hint] at how sensitizing concepts, long left silent, may murmur during coding and analysus”. She recalls how writing a memo encouraged her to look back at ideas presented in pivotal texts that she had read earlier in her career, and thereby committed her to a new strand of thought.\nCharmaz (2014: 180) describes how she developed memos from in vivo codes that recurred throughout across the cases. She asked how the saying was applied in different contexts, its overlapping and varied meaning.\nCharmaz (2014: 183-?) encourages adoption of various writing strategies. She notes that “memo-writing requires us to tolerate ambiguity”, which is inherent in the “discovery phase” of research, in which she considers memo-writing to be a part. She encourages adotion of clustering and freewriting techniques to help get the ball rolling (she refers to these as “pre-writing techniques”).\nSaldaña (2016) dedicates Chapter 2 to “writing analytic memos”. Saldaña (2016: 44) notes that codes are nothing more than labels until they are analyzed, and remarks that memo-writing is a stage in the process of getting beyond the data. He refers to Stake (1995: 19), who mused that “Good research is not about good methods as much as it is about good thinking”, and in keeping with Charmaz’s (2014) account of memo-writing, memos are tools for doing good thinking.\nSaldaña (2016: 44-45) channels Charmaz in saying that all memos are analytic memos. I identify with his discomfort in writing according to a pre-defined category of memos, and his preference to categorizing them after writing the memo. He also suggests writing a title or brief description to help with sorting.\nHowever, like Charmaz, Saldaña does differentiate between memos and “field notes”, which are synonymous with Charmaz’s journal entries. According to Saldaña (2016: 45), Field notes are the researcher’s personal and subjective responses to social actions encountered during data collection.\nSaldaña (2016: 53-54) reports on O’Connor’s (2007: 8) conceptualization of contemplation of qualitative data as refraction.\n\nThis perspective acknowledes the mirrored reality and the researcher’s lens as dimpled and broken, obscured in places, operating as a concave or at other times a convex lens. As such, it throws unexpected and distorted images back. It does not imitate what looks into the mirror but deliberately highlights some things and obscures others. It is deliciously … unpredictable in terms of what might be revealed and what might remain hidden.\n\nOther analogies include that by Stern (2007: 119): If data are the building blocks of the developing theory, memos are the mortar”, and by Birks and Mills (2015: 40) who consider memos as the “lubricant” of the analytic machine, and “a series of snapshots that chronicle your study experience”.\nSee Montgomery and Bailey (2007) and McGrath (2021) for more on the distinction between memos and field notes, including detailed examples of these kinds of writing in action." }, { - "objectID": "notes/maelstrom-readings.html#bergeron2018", - "href": "notes/maelstrom-readings.html#bergeron2018", - "title": "Maelstrom reading notes", - "section": "Bergeron et al. (2018)", - "text": "Bergeron et al. (2018)\nThe authors reference the drive for efficiency as a motivating factor that drives open data:\n\nHowever, many cohort databases remain under-exploited. To address this issue and speed up discovery, it is essential to offer timely access to cohort data and samples.\n\nHowever the paper is actually about the need for better and more publicly accessible documentation about data.\nThe authors state that catalogues exist to promote discoverability of data and samples and to answer the data documentation needs of individual studies.\nThey draw attention to the importance of catalogues in research networks (analyzing data across studies), which establish portals that document “summary statistics on study subjects, such as the number of participants presenting specific characteristics (e.g. diseases or exposures)”.\nThe authors outline several challenges that inhibit or limit the potential value of catalogues:\n\nThe quality of a catalogue directly depends on the quality and comprehensiveness of the study-specific information documented. But, maintaining and providing access to understandable and comprehensive documentation to external users can be challenging for cohort investigators, and require resources not always available, particularly for the very small or long-established studies. In addition, the technical work required to build and maintain a catalogue is particularly demanding. For example, gathering comprehensive and comparable information on study designs necessitates the implementation of rigorous procedures and working in close collaboration with study investigators. Manual classification of variables is also a long and a tedious process prone to human error. Moreover, the information collected needs to be regularly revised to update metadata with new data collections. These challenges, among others, can lead to the creation of catalogues with partial or disparate information across studies, documenting limited subsets of variables (e.g. only information collected at baseline) or including only studies with data dictionaries available in a specific language or format.\n\nBullet point summary:\n\nA catalogue’s quality depends on the quality and comprehensiveness of documentation provided by individual studies\nCohort investigators, i.e. leaders of individual studies, are under-equipped to provide such comprehensive documentation\n\nDo they just need material support? Or also guidance on how to do it, what factors to account for, etc?\n\nTechnical work for building and maintaining a catalogue is demanding\n\nI’m not sure if they example they provide to illustrate these challenges aligns with what I would call “technical work”; they refer to precise and detailed documentation in direct consultation with individual study maintainers, and I suppose the discussion about and documentation of methodological details is technical in that it corresponds with the work that was actually done on the ground, using data collection and processing instruments\n\nClassification of variables is a long and tedious process\n\nWhat makes it long and tedious? This isn’t really specified\nThey recite that this is prone to human error, but I wonder what successful or error-ful (?) outcomes would look like and how they would differ\n\nThe information needs to be regularly revised and updated\n\nThe authors recommendations to resolve these concerns:\n\nHowever, to truly optimize usage of available data and leverage scientific discovery, implementation of high quality metadata catalogues is essential. It is thus important to establish rigorous standard operating procedures when developing a catalogue, obtain sufficient financial support to implement and maintain it over time, and where possible, ensure compatibility with other existing catalogues.\n\nBullet point summary:\n\nestablish rigorous standard operating procedures when developing a catalogue\nobtain sufficient financial support to implement and maintain it over time\nwhere possible, ensure compatibility with other existing catalogues" + "objectID": "notes/methodology-notes.html#preliminary-analyses", + "href": "notes/methodology-notes.html#preliminary-analyses", + "title": "Methodology notes", + "section": "Preliminary analyses", + "text": "Preliminary analyses\nYin (2014: 135-136 5) identifies various strategies for analyzing case study evidence.\n\nA helpful starting point is to “play” with your data. You are searching for patterns, insights, or concepts that seem promising. (Yin 2014: 135)\n\nCiting Miles and Huberman (1994), Yin (2014) lists a few strategies at this playful stage:\n\nJuxtaposing data from different interviews\nPutting information into different arrays\nMaking a matrix of categories and placing the evidence within them\nTabulating the frequency of different events\nPutting information in chronological order or using some other temporal scheme\n\nYin (2014: 135) also emphasizes memo-writing as a core strategy at this stage, citing Corbin and Strauss (2014). These memos should include hints, clues and suggestions that simply put into writing any preliminary interpretation, essentially conceptualizing your data. He uses the specific example of shower thoughts.\n\nSaldaña (2016) dedicates chapter 6 to “post-coding and pre-writing transitions”. He frames these as a series of strategies to help crystallize the analytical work and springboard into written documents and reports.\n\nTop-10 lists\nSaldaña (2016: 274-275) suggests coming up with a list of the ten top quotes or passages as a potentially useful “focusing strategy”. Identify the passages (no longer than half a page each) that strike me as the most vivid and/or representational of my study. He suggests reflecting on the content of these items and arranging them in various orders to discover different ways of structuring or outlining the write-up of the research story. He provides some examples of orders to consider:\n\nchronologically\nhierarchically\ntelescopically\nepisodically\nnarratively\nfrom the expository to the dramatic\nfrom the mundane to the insightful\nfrom the smallest detail to the bigger picture\n\n\n\nThe study’s “trinity”\nSaldaña (2016: 275-276) suggests identifying the three (and only three) major codes, categories, themes and/or concepts generated thus far that strike me or stand out in my study. He suggests identifying which one is dominant, how does this status relate to or impact the other codes or concepts, and generally trace the relationships between these ideas. He suggests plotting them as a venn diagram to identify what aspects overlap across items, and to label those areas of overlap — although he does not mention this explicitly, I imagine these points of overlap represent the synthesis of new emergent ideas.\n\n\nCodeweaving\nSaldaña (2016: 276) addresses codeweaving as a viable strategy, but I don’t actually think of it as that useful for my purposes. Seems a but contrived and has lots of potential to be over-engineered.\n\n\nFrom coding to theorizing\nFor Saldaña (2016: 278), the stage at which he finds theories emerging in his mind is when he starts coming up with categories of categories. At this point, a level of abstraction occurs that transcends the particulars of a study, enabling generalizable transfer to other comparable contexts.\nSaldaña (2016: 278-279) identifies a few structures through which these categories of categories might emerge:\n\nSuperordindate and Subordinate Arrangements: Arrange categories as an outline, which suggests discrete linearity and classification. Supercategories and subcategories are “ranked” with numbers or capital letters.\nTaxonomy: Categories and their subcategories are grouped but without any inferred hierarchy; each category seems to have equal weight.\nHierarchy: Categories are ordered from most to least in some manner, i.e. frequency, importance, impact, etc.\nOverlap: Some categories share particular features with others while retaining their unique properties.\nSequential Order: The action suggested by categories progresses in a linear manner.\nConcurrency: Two or more categories operate simultaneously to influence and affect a third.\nDomino Effects: Categories cascade forward in multiple pathways.\nNetworks: Categories interact and interplay in complex pathways to suggest interrelationship.\n\nThe arrows connecting categories are meaningful in their own right. Saldaña (2016: 280) references Urquhart (2013) who states that category relationships are necessary to develop assertions, propositions, hypotheses and theories. He suggests inserting words or phrases between categories that plausibly establishes their connections, as suggested by the data and analytical memos. Saldaña (2016: 280-281) lists several possible connectors:\n\naccelerates\ncontributes toward\ndepends on the types of\ndrives\nfacilitates\nharnesses\nincreases\nincreases the difficulty of\nis affected by\nis essential for\nis necessary for\nprovides\nreconciles\nreduces\nresults in\nresults in achieving\ntriggers\nvaries according to\nwill help to\n\nMoreover, Saldaña (2016: 281) suggests that if you end up with categories as nouns or noun phrases, it could be helpful to transform them into gerund phrases. This will help get a better sense of process and action between catagories.\n\n\nFindings “at a glance”\nFollowing Henwood and Pidgeon (2003), Saldaña (2016: 283) suggests creating a tri-column chart that outlines the findings and the sources of evidence and reasoning that underlie them. See the specific page for a good example." }, { - "objectID": "notes/maelstrom-readings.html#bergeron2021", - "href": "notes/maelstrom-readings.html#bergeron2021", - "title": "Maelstrom reading notes", - "section": "Bergeron et al. (2021)", - "text": "Bergeron et al. (2021)\nIdentifies several registries of relevant cohorts, but notes that they face challenges getting the data together. Namely, issues concerning institutional policies concerning data-sharing, lack of open access to cohort data and to documentation about the data, the data’s complexity which makes it difficult to harmonize across studies, and lack of access to funding, secure data environments, and specialized expertise and resources.\nThe Research Advancement through Cohort Cataloguing and Harmonization (ReACH) initiative was establihed in collaboration with Maelstrom to overcome some of these barriers in the context of Developmental Origins of Health and Disease (DOHaD) research.\nThe authors briefly summarize some projects that rely on ReACH data, and provide a more comprehensive table of ongoing and unpublished work.\nIn the supplementary materials, the authors also include an illustrative example specific tasks, decisisions and actions that one might got through when using ReACH data. It is a broad-level but fairly sober account of how one would navigate the catalogue and engage with collaborators." + "objectID": "notes/methodology-notes.html#the-constant-comparative-method", + "href": "notes/methodology-notes.html#the-constant-comparative-method", + "title": "Methodology notes", + "section": "The constant comparative method", + "text": "The constant comparative method\nThe constant comparative method is based on action codes, similar to what Saldaña (2016) refers to as process codes. According to Charmaz (2000: 515):\n\nThe constant comparative method of grounded theory means (a) comparing different people (such as their views, situations, actions, accounts, and experiences), (b) comparing data from the same individuals with themselves at different points in time, (c) comparing incident with incident, (d) comparing data with categori, and (e) comparing categories with other categories.\n\nMy initial impression is that this is very well suited for Stake’s (2006) multicase study framework, specifically with regards to his notion of the case-quintain dilemma. It also seems very well suited for analysis of situational meaning-making, as per Suchman (1987), Lave and Wenger (1991), Knorr Cetina (2001) and symbolic interactionism at large." }, { - "objectID": "notes/maelstrom-readings.html#wey2024", - "href": "notes/maelstrom-readings.html#wey2024", - "title": "Maelstrom reading notes", - "section": "Wey and Fortier (2024)", - "text": "Wey and Fortier (2024)\nAn overview of the harmonization procedires applied in CanPath and MINDMAP. Authored by two members of the Maelstrom team, but no one from these initiatives.\nAt first, comes across as another broad-level overview of processes. But, as is elicited in the conslusion, the paper highlights some subtle divergent approaches, some of which are relevant to my project and I picked up here.\nInteresting bit about use of centralized systems or systems that are more conducive to end-users’ individual workflows.\n\nIn both illustrative projects, information about participating studies and data collected was gathered and made available on the central data server. In the CanPath BL-HRFQ, harmonization documentation and processing scripts were held centrally and only updated by Maelstrom Research. In MINDMAP, because multiple groups simultaneously worked on generating harmonization processing scripts for different areas of information, working versions of the DataSchema and processing scripts were held on a GitHub, allowing for better version control and dynamic updating of scripts by multiple remote data harmonizers.\n\nMore about the balance being struck at MINDMAP between institutional and popular tech and documentation platforms:\n\nIn the MINDMAP project, R markdown documents with applied harmonization scripts (Figure 13.1b) and any comments were preserved in the GitHub repository, which is also publicly accessible. Furthermore, summary statistics for MINDMAP variables are only available to approved data users through the secure server, as harmonized datasets are only intended for use within the MINDMAP network. Overviews of the harmonization process and outputs have also been published as open-access, peer-reviewed articles for both projects (Fortier et al. 2019; Wey et al. 2021).\n\nBit about existing collaborative ties making things much easier:\n\nthe prospective coordination and unu sually high ability for the harmonization team (Maelstrom Research) to work directly with study data managers on recently collected and documented study data resulted in high standardization and ability to resolve questions.\n\nData curation as an explicitly creative task, involving impactful decisions. Interesting that this was unexpected enough to warrant inclusion as a key challenge:\n\nIn MINDMAP, these differences were clearly documented in comments about harmonization for each source dataset to allow researchers to decide how to use the harmonized variable. In-depth exploration of the best statistical methods to harmonize these types of measures to maintain integrity of content while minimizing loss of information and methodological bias are important and active areas of research (Griffith et al. 2016; Van den Heuvel and Griffith 2016). The approach taken in this case was to apply simpler harmonization meth ods (i.e. rescaling) and document the transformation, leaving investigators the flexibility to further explore and analyze the harmonized datasets as appropriate for their research questions.\n\nAs with the above excerpt, documentation was considered as a viable mode of resolving or accounting for significant discrepancies:\n\nThe papers describing the harmonization projects attempt to highlight these considerations, for example, providing some comparison of demographics of the harmonized populations against the general populations from which they were drawn (Dummer et al. 2018; Fortier et al. 2019; Wey et al. 2021), listing sources of study-specific heterogeneity in the harmonized datasets to consider (Fortier et al. 2019), and pointing users to individual study documentation where more information on weights to use for analysis should be con sidered (e.g. Wey et al. 2021).\n\nThe bit about use of github was rationalized as a way of facilitating collaboration across insitutional boundaries. They refer to R and RMarkdown being open standards as the main reason, but I wonder if a part of it is that GitHub, being a non-institutional platform, was easier to use from an oboarding perspective:\n\nIn MINDMAP, due to the need for multiple international groups to work simultaneously and flexibly on the harmonization processing and frequently evolving versions of study-specific harmonization scripts, scripts for harmonization processing were written and applied entirely through R markdown in an RStudio interface, and the DataSchema and R markdown versions were maintained and frequently updated in a GitHub repository.\n\nPerhaps ironically, the private company and platform may have been used due to the strength of pan-european collaborative ties that institutions may not be able to keep up with. Whereas in Canada, with centralized project-oriented work, it may be much easier to enforce adoption of centralized tooling. This is just speculation." + "objectID": "notes/methodology-notes.html#situational-analysis", + "href": "notes/methodology-notes.html#situational-analysis", + "title": "Methodology notes", + "section": "Situational analysis", + "text": "Situational analysis\nSituational analysis originates from Strauss’s social worlds/arenas/negotiations framework. From Clarke (2003: 554):\n\nBuilding on and extending Strauss’s work, situational analyses offer three main cartographic approaches:\n\nsituational maps that lay out the major human, nonhuman, discursive, and other elements in the research situation of concern and provoke analyses of relations among them;\nsocial worlds/arenas maps that lay out the collective actors, key nonhuman elements, and the arena(s) of commitment within which they are engaged in ongoing negotiations, or mesolevel interpretations of the situation; and\npositional maps that lay out the major positions taken, and not taken, in the data vis-à-vis particular discursive axes of variation and difference, con cern, and controversy surrounding complicated issues in the situation.\n\n\nClarke (2003: 560) identifies the main purpose of situational data as a way of “opening up” the data, figuing out where and how to enter:\n\nAlthough they may do so, a major and perhaps the major use for them is “opening up” the data –— interrogating them in fresh ways. As researchers, we constantly confront the problem of “where and how to enter.” Doing situational analyses offers new paths into the full array of data sources and lays out in various ways what you have to date. These approaches should be considered analytic exercises — constituting an ongoing research “work out” of sorts—well into the research trajectory. Their most important outcome is provoking the researcher to analyze more deeply.\n\nShe emphasizes that this is meant to stimulate thinking, and should always be paired with comprehensive memoing before, during and after situation mapping excercises.\nA key feature that I think is invaluable is the ability to uncover the sites of silence, or the things that I as a researcher suspect are there but are not readily visible in my evidence. Situational analysis is useful for drawing out the thousand pound gorillas in the room that no one wants to talk about, and is therefore important for identifying things to address during continual data collection, as is one of the (often ignored) central pillars of grounded theory:\n\nThe fourth and last caveat is perhaps the most radical. As trained scholars in our varied fields, usually with some theoretical background, we may also suspect that certain things may be going on that have not yet explicitly appeared in our data. As ethically accountable researchers, I believe we need to attempt to articulate what we see as the sites of silence in our data. What seems present but unarticulated? What thousand-pound gorillas are sitting around in our situations of concern that nobody has bothered to mention as yet (Zerubavel 2002)? Why not? How might we pursue these sites of silence and ask about the gorillas without putting words in the mouths of our participants? These are very important directions for theoretical sampling. That is, the usefulness of the approaches elucidated here consists partly in helping the researcher think systematically through the design of research, espe cially decisions regarding future data to collect.\n\nThe process is remarkably similar to the brainstorming excercise I did with Costis one time. Starts by articulating the actors involved, their roles and relationships, the things they do, the things they make.\n\nThe goal here is to lay out as best one can all the human and nonhuman elements in the situation of concern of the research broadly conceived. In the Meadian sense, the questions are: Who and what are in this situation? Who and what matters in this situation? What elements “make a difference” in this situation?\n\nAfter jotting these down on a canvas or whiteboard, arrange them into more concrete categories. In each category, refer to examples or specific instances. Concepts can occur in multiple categories. By arranging these concepts, grouping them thematically, spatially, and through relationships and associations with arrows or lines (presumably, with labels indicating the nature of these relationships), this provides the researcher with a viable way of unlocking new pathways to think through their data. I imagine this will be especially helpful when arranging or re-arranging the coding tree.\nClarke (2003: 569-570) also dedicates a few paragraphs to relational forms of analysis using this technique:\n\nRelations among the various elements are key. You might not think to ask about certain relations, but if you do what I think of as quick and dirty relational analyses with the situational map, they can be revealing. The procedure here is to take each element in turn and think about it in relation to each other element on the map. One does this by circling one element and mentally or literally drawing lines, one at a time, between it and every other ele ment on the map and specifying the nature of the relationship by describing the nature of that line.\n\n\nYou could highlight (in blue perhaps) that organization’s perspectives on all the other actors to see which actors are attended to and which are not, as well as the actual contents of the organization’s discourses on its “others.” Silences can thus be made to speak.\n\nClarke (2003: 569) On how to generate memos from situational analysis:\n\nEach map in turn should lead to a memo about the relations diagrammed. At early stages of analysis, such memos should be partial and tentative, full of questions to be answered about the nature and range of particular sets of social relations, rather than being answers in and of themselves.\n\nClarke (2003: 570) addresses the energy or mood required to do this kind of work, which is reminiscient of Yin’s (2014: 73-74) description of the “mental and emotional exhaustion at the end of each fieldwork day, due to the depletion of ‘analytic energy’ associated with being attention on your toes.” Interesting that the mention of “freshness” is brought up in the context of Glaser’s tabla rasa approach to grounded theory.\n\nAs a practical matter, doing the situational map and then the relational analyses it organizes can be tiring and/or anxiety producing and/or elating. Work on it until you feel stale and then take a break. This is not the same order of work as entering bibliographic citations. The fresher you are, the more you can usually see. Glaser (1978: 18-35) cautions against prematurely discussing emergent ideas –— that we might not necessarily benefit from talking about everything right away but rather from reflection — and memoing. I strongly agree, especially about early even if quick and dirty memoing. But we all must find our own ways of working best. For most, the work of this map occurs over time and through multiple efforts and multiple memos.\n\nClarke (2003) refers to Shim (2000) as an exemplary case of situational analysis in action." }, { - "objectID": "notes/maelstrom-readings.html#doiron2013a", - "href": "notes/maelstrom-readings.html#doiron2013a", - "title": "Maelstrom reading notes", - "section": "Doiron, Raina, and Fortier (2013)", - "text": "Doiron, Raina, and Fortier (2013)\nThis paper summarizes what was discussed at a workshop bringing together stakeholders who would contribute to two large data harmonization initiatives: the Canadian Longitudinal Study on Aging (CLSA) and the Canadian Partnership for Tomorrow Project (CPTP). It is therefore representative of plans and challenges that were held at an early stage when collaborations were being established.\nThe authors identify series of reasons for linking data, which I summarize here:\n\nMaximizing potential of disparate information resources\n\n\nenriching study datasets with additional data not being collected directly from study par ticipants\noffer vital information on health outcomes of participants\nvalidate self-reported information\n\n\nDrawing maximum value from data produced from public expenditure\n\n\noffers a cost-effective means to maximize the use of existing publicly funded data collections\n\n\nDevelop interdisciplinary collaborative networks\n\n\nby combining a wide range of risk factors, disease endpoints, and relevant socio-economic and biological measurements at a population level, linkage lays the groundwork for multidisciplinary health-research initiatives, which allow the exploration of new hypotheses not foreseeable using independent datasets\n\n\nEstablish long-lasting infrastructure and instill a collaborative culture\n\n\nLast, a coordinated pan-Canadian cohort-to-administrative linked database would establish legacy research infrastructures that will better equip the next generation of researchers across the country\n\nThe authors use the term “data linkage”:\n\nData linkage is “the bringing together from two or more different sources, data that relates to the same individual, family, place or event”. When linking data at the individual level, a common identifier (or a combination of identifiers) such as a personal health number, date of birth, place of residence, or sex, is used to combine data related to the same person but found in separate databases. Data linkage has been used in a number of research fields but is an especially valuable tool for health research given the large amount of relevant information collected by institutions such as governments, hospitals, clinics, health authorities, and research groups that can then be matched to data collected directly from consenting individuals participating in health research.\n\nThis is distinct from harmonization in that it is not meant to combine data with similar scope and schematic structure, but rather to relate information collected under various domains so that they could be more easily queried in tandem. I imagine this as reminiscient of establishing links between tables in a relational database.\nThe authors identify the open-endedness of the linked data as a unique challenge, without elaborating on this point:\n\nCLSA/CPTP-to-AHD linkage also poses unique challenges in that, in contrast to more traditional requests to link data to answer one-off research questions, it aims to establish a rich data repository that will allow investigators to answer a multitude of research questions over time.\n\nThe workshop participants established a 5-point plan:\n\nbuild strong collaborative relationships between stakeholders involved in data sharing (e.g., researchers, data custodians, and privacy commissioners);\nidentify an entity which could provide overall leadership as well as individual “champions” within each province;\nfind adequate and long-term resources and funding;\nclarify data linkage and data-sharing models and develop a common framework within which the data linkage process takes place; and\ndevelop a pilot project making use of a limited number of linked variables from participating provinces\n\nThe second point, about identifying “champions”, is kind of interesting, and I’d like to know more about what qualities these people were expcted to have, their domains of expertise, their collaborative/soft or technical skills, and how this plays into access to funds and general governance structures\nNeed to look at Roos, Menec, and Currie (2004), which they cite in the conclusion, specifically with reference to the aspiration to develop “information rich environments”. Seems like it is a primary source for the background on linked data in Manitoba and Australia." + "objectID": "notes/methodology-notes.html#various-other-analytical-strategies-and-techniques", + "href": "notes/methodology-notes.html#various-other-analytical-strategies-and-techniques", + "title": "Methodology notes", + "section": "Various other analytical strategies and techniques", + "text": "Various other analytical strategies and techniques\nYin (2014: 136-142) describes four general analytical strategies:\n\nRelying on theoretical propositions\nWorking your data from the “ground up”\nDeveloping a case description\nExamining plausible rival explanations\n\nYin (2014: 142-168) then goes on to describe five analytical techniques:7\n\nPattern matching\nExplanation building\nTime-series analysis\nLogic models\nCross-case synthesis\n\nRyan and Bernard (2000) describe various analysis techniques for analyzing textual elicitations in structured and codified ways." }, { - "objectID": "notes/maelstrom-readings.html#fortier2023", - "href": "notes/maelstrom-readings.html#fortier2023", - "title": "Maelstrom reading notes", - "section": "Fortier et al. (2023)", - "text": "Fortier et al. (2023)\nRelates harmonization to the FAIR principles, which has not really been featured much in arguments for harmonization in prior Maelstrom papers. Specifically, this paper frames harmonization as a necessary additional condition that enables FAIR data to be made useful; FAIR is deemed not enough.\n\nIn the following paper, we aim to provide an overview of the logistics and key ele ments to be considered from the inception to the end of collabo rative epidemiologic projects requiring harmonizing existing data.\n\nInteresting acronym/framework for defining research questions:\n\nThe research questions addressed and research plan proposed by harmonization initiatives need to be Feasible, Interesting, Novel, Ethical, and Relevant (FINER). (Cummings, Browner, and Hulley 2013)\n\nTable 1 lists examples of questions that could be addressed to help delineate analytical approach, practical requirements, and operations of a harmonization initiative. These are all questions that look toward or project specific goals, which then inform the strategies through which they may be achieved.\nThe supplementary materials include very detailed information about specific practices and objectives pertaining to the REACH initiative. However, it’s unclear how well this reflects any specific challenges experienced. In other words, I was hoping for something more candid.\nMoreover, this paper is also based on survey responses from 20 harmonization initiatives, but neither the findings resulting from the analysis, nor the data, are referenced or included. Is this the same as the one that informed Fortier et al. (2017)?\nLooming in the background of this paper is the DCC lifecycle model. However they do not cite DCC, or the field of digital curation in general. The DCC lifecycle model has always been presented as a guideline, sort of divorced from practical experience, or at least that’s how I’ve always understood it. Basically, and literally, a model for expected behaviours and interstitial outcomes. I think it would be interesting to explore (a) why the authors perceived need to present a model and (b) how they arrived at the phases and principles that are included in it. Is this tacit or common-sense thinking? Or was this directly informed by my concrete thinking from the field of digital curation?\nI should brush up on more recent work regarding the DCC, and specifically, critiques of it. Through a quick search, a few papers seem directly relevant:\n\nCox and Tam (2018)\nChoudhury, Huang, and Palmer (2020)\nRhee (2024)\n\nI think Cox and Tam (2018) may be especially relevant as a critique of the lifecycle metephor in a general sense. From the abstract, it seems like they identify various downsides pertaining to lifecycle models, specifically that they “mask various aspects of the complexity of research, constructing it as highly purposive, serial, uni-directional and occurring in a somewhat closed system.” I’m not sure how explicit the connection is, but I sense this ties into the general paradigm shift toward greater recognition of curation “in the wild”, as per Dallas (2016)." + "objectID": "notes/methodology-notes.html#qda-software-and-tooling", + "href": "notes/methodology-notes.html#qda-software-and-tooling", + "title": "Methodology notes", + "section": "QDA software and tooling", + "text": "QDA software and tooling\nWeitzman (2000) provides an overview of software and qualitative research, including a minihistory up to the year 2000 when the chapter was published.\nDescribing the first programs specifically designed for analysis of qualitative data, Weitzman (2000: 804) writes:\n\nEarly programs like QUALOG and the first versions of NUDIST reflected the state of computing at that time. Researchers typically accomplished the coding of texts (tagging chunks of texts with labels — codes — that indicate the conceptual categories the researcher wants to sort them into) by typing in line numbers and code names at a command prompt, and there was little or no facility for memoing or other annotation or markup of text.8 In comparison with marking up text with coloured pencils, this felt awkward to many researchers. And computer support for the analysis of video or audio data was at best a fantasy.\n\nThis history if followed by a sober account of what software can and can not do in qualitative research, as well as affirmation and dismissed of hopes and fears. Very reminiscient of Huggett (2018)." }, { - "objectID": "notes/maelstrom-readings.html#gaye2014", - "href": "notes/maelstrom-readings.html#gaye2014", - "title": "Maelstrom reading notes", - "section": "Gaye et al. (2014)", - "text": "Gaye et al. (2014)\nIntroduces DataShield.\nFrames DataShield as a technical fix to administrative problems:\n\nMany technical and policy measures can be enacted to render data sharing more secure from a governance per spective and less likely to result in loss of intellectual prop erty. For example, data owners might restrict data release to aggregate statistics alone, or may limit the number of variables that individual researchers might access for speci fied purposes. Alternatively, secure analysis centres, such as the ESRC Secure Data Service and SAIL represent major informatics infrastructures that can provide a safe haven for remote or local analysis/linkage of data from selected sources while preventing researchers from down loading the original data themselves. However, to comple ment pre-existing solutions to the important challenges now faced, the DataSHIELD consortium has developed a flexible new way to comprehensively analyse individual level data collected across several studies or sources while keeping the original data strictly secure. As a technology, DataSHIELD uses distributed computing and parallelized analysis to enable full joint analysis of individual-level data from several sources, e.g. research projects or health or administrative data—without the need for those data to move, or even be seen, outside the study where they usually reside. Crucially, because it does not require underpin ning by a major informatics infrastructure and because it is based on non-commercial open source software, it is both locally implementable and very cost effective.\n\nAdds a social/collaborative element to earlier arguments about the challenges inherent of prospective harmonization, highlighting a need for engagement with individual studies (either through direct or peripheral participation) to conduct research that was not initially planned for:\n\nUnfortunately, both [study-level metadata] SLMA and [individual-level metadata] ILMA present significant problems Because SLMA com bines analytical results (e.g. means, odds ratios, regression coefficients) produced ahead of time by the contributing studies, it can be very inflexible: only the pre-planned analyses undertaken by all the studies can be converted into joint results across all studies combined. Any additional analyses must be requested post hoc. This hinders exploratory analysis for example the investigation of sub-groups, or interactions between key variables.\n\nProvides a detailed overview of how DataShield was implemented for HOP (Healthy Obesity Project), including the code used to generate specific figures and analyses. Hoever it does not really describe or reflect upon the processes through which the code was developed.\nThe authors highlight the fact that certain analytical approaches are not possible using DataShield, especially analysis that visualize individual data points. It’s unclear how they enforce this, or whether it’s an implicit limitation based on the data that DataShield participants provide.\n\nBecause in DataSHIELD potentially disclosive com mands are not allowed, some analyses that are possible in standard R are not enabled. In essence, there are two classes of limitation on potential DataSHIELD functional ity: (i) absolute limitations which require an analysis that can only be undertaken by enabling one of the functional ities (e.g. visualizing individual data points) that is explicitly blocked as a fundamental element of the DataSHIELD philosophy. For example, this would be the case for a standard scatter plot. Such limitations can never be circumvented and so alternatives (e.g. contour and heat map plots) are enabled which convey similar information but without disclosing individual data points; (ii) current limitations which are functions or models that we believe are implementable but we have not, as yet, under taken or completed the development work required. As examples, these latter include generalized linear mixed model (including multi-level modelling) and Cox regression.\n\nThe authors list numerous other limitations and challenges. Some have to do with what kinds of data DataShield can handle (something about horizontal and vertical that I do not yet fully understand). Other challenges include the need for data to be harmonized, and having to deal with governance concerns.\nNotably, the first challenge mentioned seems to contradict the statement earlier on (and made by Doiron et al. (2013)) that this is relatively easy to set up. The authors acknowledge the fact that coding for analysis using DataShield has a steep learning curve and requires some pre-planning to enable results from satellite computers to be properly combined. Their mitigation is to black-box these concerns by implementing simpler client-side functions that mask the more complex behaviours (and presumably translate error messages in ways that users can understand and act to resolve!).\n\nDespite its potential utility, implementation of DataSHIELD involves significant challenges. First, although set-up is fundamentally straightforward, application involves a relatively steep learning curve because the command structure is complex: it demands specification of the analysis to be undertaken, the studies to use and how to combine the results. In mitigation, most complex serverside functions are now called using simpler client-side functions and we are working on a menu-driven implementation.\n\nAlso interesting that they note how there may be unanticipated problems, either accidental or malicious, and their way of mitigating against this is to log all commands:\n\nFifth, despite the care taken to set up DataSHIELD so that it works properly and is non-disclosive, it is possible that unanticipated prob lems (accidental or malicious) may arise. In order to iden tify, describe and rectify any errors or loopholes that emerge and in order to identify deliberate miscreants, all commands issued on the client server and enacted on each data server are permanently logged.\n\nThis is even more interesting in light of their continuous reference to “data.care”, which they do not address in depth, but which seems to have been a scandal involving unauthorized release of personal health data used in research.\nThe authors add an additional caveat concerning the need to ensure that the data are cleaned in advance.\n\nBut, to be pragmatic, many of the routinely collected healthcare and administra tive databases will have to undergo substantial evolution before their quality and consistency are such that they can directly be used in high-quality research without exten sive preparatory work. By its very nature, such preparation—which typically includes data cleaning and data harmonization—cannot usually be undertaken in DataSHIELD, because it involves investigating discrepan cies and/or extreme results in individual data subjects: the precise functionality that DataSHIELD is designed to block. Such work must therefore be undertaken ahead of time by the data generators themselves—and this is de manding of time, resources and expertise that — at present — many administrative data providers may well be unwilling and/or unable to provide. That said, if the widespread us ability of such data is viewed as being of high priority, the required resources could be forthcoming.\n\nThis corresponds with another limitation identified earlier, namely with regards to identifying duplicate individual records across jurisdictional boundaries (which involves assumptions regarding nationality and identify – one of those weird myths that programmers can’t seem to let go!):\n\nSo far DataSHIELD has been applied in settings where individual participants in different studies are from different countries or from different regions so it is unlikely that any one person will appear in more than one source. However, going forward, that cannot al ways be assumed. We have therefore been consider ing approaches to identify and correct this problem based on probabilistic record linkage. In the genetic setting 48 the BioPIN provides an alternative solution. Ongoing work is required.\n\nNote the last line of the prior block quote regarding data cleaning:\n\nThat said, if the widespread us ability of such data is viewed as being of high priority, the required resources could be forthcoming.\n\nThis seems like a thread worth tugging at!" + "objectID": "notes/methodology-notes.html#writing-techniques", + "href": "notes/methodology-notes.html#writing-techniques", + "title": "Methodology notes", + "section": "Writing techniques", + "text": "Writing techniques\n\nAnalytical storylining\nFollowing Charmaz (2010: 201), Saldaña (2016: 287) suggests using processual words that suggest the unfolding of events through time that “reproduce the tempo and mood of the [researched] experience”. However, he cautions that this is not suitable for every situation, and to be aware of how this framing differs from positivist notions of cause and effect.\n\n\nOne thing at a time\nSaldaña (2016: 287) suggests documenting each category, theme or concept one at a time as a wat of disentangling very interrelated ideas. This helps maintain focused as a writer, and may help maintain a more focused reading experience too.\n\n\nWriting about methods\nSaldaña (2016: 284-286) addresses some common tendencies and challenges with regards to writing about qualitative coding practices:\n\nResearchers provide brief storied accounts of what happened “backstage” at computer terminals and during intensive work sessions. After a description of the participants and the particular data collected for a study, descriptions of coding and analysis procedures generally include: references to the literature that guided the analytic work; qualitative data organization and management strategies for the project; the particular coding and data analytic methods employed and their general outcomes; and the types of CAQDAS programs and functions used. Some authors may include accompanying tables or figures that illustrate the codes or major categories constructed. Collaborative projects usually explain team coding processes and procedures for reaching intercoder agreement or con-sensus. Some authors may also provide brief confessional anecdotes that highlight any analytic dilemmas they may have encountered.\n\nSaldaña (2016: 286) then suggests emphasizing for readers through introductory phrases and italicized text the major outcomes of the analysis to help provide guidance for readers to grasp the headlines of your research story. For example:\n\n“The three major categories constructed from transcript analysis are …”\n“The core category that led to the groundedtheory is ..,” “The key assertion of this study is …”\n“A theory I propose is …”" }, { - "objectID": "notes/maelstrom-readings.html#wolfson2010", - "href": "notes/maelstrom-readings.html#wolfson2010", - "title": "Maelstrom reading notes", - "section": "Wolfson et al. (2010)", - "text": "Wolfson et al. (2010)\nx" + "objectID": "notes/methodology-notes.html#footnotes", + "href": "notes/methodology-notes.html#footnotes", + "title": "Methodology notes", + "section": "Footnotes", + "text": "Footnotes\n\n\nThe term refers to a medieval jousting target: see https://en.wikipedia.org/wiki/Quintain_(jousting)↩︎\nThough Yin (2014: 40-444) is dismissive of such use of the term “sample” since he sees case study research as only generalizable to similar situations, and not to a general population from which a sample is typically said to be drawn. I agree with this focus on concrete situations over Stake’s prioritization of theory-building as an end unto itself.↩︎\nCharmaz uses the term “giving voice” in this specific context. I’m not sure if this is meant to represent Strauss and Corbin’s attitude, and whether this is an accurate representation on their views, but in my mind this should be framed as elevating, amplifying or re-articulating respondents’ voices (and this is a tenet of constructivist grounded theory in general, which derives from Charmaz). My take diverges from the position that we “give voice” to respondents in that it acknowledges (1) that the voices are already there, (2) that respondents are in fact giving us their voices, and (3) that the researcher plays an active editorial role, transforming the respondents’ elicitations into a format that is more amenable to analysis.↩︎\nVery much in line with the pragmatist turn of the late ’90s and early ’00s, as also documented by Lucas (2019: 54-57) in the context of archaeological theory, vis-a-vis positivism, postmodernism, and settling on a middle ground between them.↩︎\nThis seems reminiscient of procedures planned for WP1 and WP2 from my FWO grant application.↩︎\nThis is also very reminiscient of Nicolini (2009).↩︎\nI wonder: would Abbott (2004) call these heuristics?↩︎\nThis caught my eye since its the same approach as that adopted by qc!↩︎" }, { - "objectID": "interview-protocol.html", - "href": "interview-protocol.html", - "title": "Interview Protocol", + "objectID": "notes.html", + "href": "notes.html", + "title": "Notes", "section": "", - "text": "Interviews may be held either in-person or through online video conference. All interviews will be held in quiet and comfortable environments, such as office spaces or conference rooms.\nI will record all in-person interviews using a SONY ICD-UX560 audio recorder to capture audio in the lossless 16 bit 44.1 kHz Linear PCM wav format, with additional audio filters to enhance playback during transcription, if necessary.\nI will also record in-person interview sessions using a GoPro Hero 4 Silver action camera, depending on participants willingness to be video recorded. Based on prior interviews with scientists about their research experiences, I found that interviewees like to show me, rather than merely tell me, about what they are working on and the means through which they engage with information systems. The camera may be leveraged to record spontaneous video records of these demonstrations and provide me with an additional rich data source for further analysis. Moreover, the camera provides an additional backup audio recording in case of data loss on the primary recording device.\nRemote interviews will be recorded using the video conferencing software’s built-in recording tools. Participants will be instructed to disable their microphones or video cameras prior to initiating recording if they have opted to not be recorded through these media. The researcher will record all media locally and refrain from using any cloud services to store or modify the records which the video conference software may provide.\nI will also record handwritten notes comprising descriptive accounts of activities and interactions when recording devices are switched off, as well as preliminary interpretations of observed behaviours and notes on things I plan to follow up on at a later time.", + "text": "Modified\n\n\nTitle\n\n\nCategories\n\n\n\n\n\n\nFeb 21, 2025\n\n\nMethodology notes\n\n\nreading, general thoughts\n\n\n\n\nFeb 21, 2025\n\n\nPotential cases\n\n\ncases, brainstorming\n\n\n\n\nJan 28, 2025\n\n\nMaelstrom reading notes\n\n\nreading\n\n\n\n\n\nNo matching items", "crumbs": [ - "Interview Protocol" + "Notes" ] }, { - "objectID": "interview-protocol.html#interview-records", - "href": "interview-protocol.html#interview-records", - "title": "Interview Protocol", + "objectID": "index.html", + "href": "index.html", + "title": "CITF-Postdoc", "section": "", - "text": "Interviews may be held either in-person or through online video conference. All interviews will be held in quiet and comfortable environments, such as office spaces or conference rooms.\nI will record all in-person interviews using a SONY ICD-UX560 audio recorder to capture audio in the lossless 16 bit 44.1 kHz Linear PCM wav format, with additional audio filters to enhance playback during transcription, if necessary.\nI will also record in-person interview sessions using a GoPro Hero 4 Silver action camera, depending on participants willingness to be video recorded. Based on prior interviews with scientists about their research experiences, I found that interviewees like to show me, rather than merely tell me, about what they are working on and the means through which they engage with information systems. The camera may be leveraged to record spontaneous video records of these demonstrations and provide me with an additional rich data source for further analysis. Moreover, the camera provides an additional backup audio recording in case of data loss on the primary recording device.\nRemote interviews will be recorded using the video conferencing software’s built-in recording tools. Participants will be instructed to disable their microphones or video cameras prior to initiating recording if they have opted to not be recorded through these media. The researcher will record all media locally and refrain from using any cloud services to store or modify the records which the video conference software may provide.\nI will also record handwritten notes comprising descriptive accounts of activities and interactions when recording devices are switched off, as well as preliminary interpretations of observed behaviours and notes on things I plan to follow up on at a later time.", - "crumbs": [ - "Interview Protocol" - ] - }, - { - "objectID": "interview-protocol.html#interview-guide", - "href": "interview-protocol.html#interview-guide", - "title": "Interview Protocol", - "section": "Interview guide", - "text": "Interview guide\nInterviews are oriented by my goal to document values and attitudes concerning data harmonization efforts, as elicited by research participants in their responses. Participants will be asked to reflect on:\n\nthe motivations for their initiatives;\nthe challenges they experience;\nhow they envision success and failure;\ntheir perceptions of their own roles and the roles of other team members and stakeholders;\nthe values that inform their decisions;\nhow the technological apparatus they set up enables them to realize their goals and values; and\nways in which they believe data-sharing could be improved.\n\nTo this end, each interview will proceed following a strategic order:\n\n1. Participants’ goals and perspectives\nFollow a life-history method to better understand participants’ professional backgrounds and their roles within their projects. The goal is to obtain information about their paths, not the rehearsed origin story.\n\nTo start, can you please tell me a little bit about your background?\nWhat is the project, and what is your role?\nHow did you find yourself in this role?\nHow has your previous experience prepared you for your role?\n\n\n\n2. Projects’ missions, purposes, motivations\nThis section is about the project in general, including its purpose, scope and value. Information about practices and procedures will be sought in a subsequent phase of the interview.\n\nWhat are the project’s goals?\nWhat makes the project unique?\nWhat is the project doing that no other similar project is doing?\nDo you consider this project as similar to any other initiatives?\nWhat are they, and in what ways are they simiar or different?\n\n\n\nWhat are the expected outcomes?\nHave you achieved these goals and outcomes?\nIf not Are you on track to achieving them?\nWhat are some challenges that the project experienced, and how have you worked to overcome them?\n\n\n\n3. Practices, procedures, relationships\nThis section asks about specific actions and interactions that the participant engages in.\n\nRoles and relationships\n\nWhat does your role entail?\nCan you provide a couple examples of things that you recently did in this capacity?\n\n\n\nWho else do you frequently rely on, and what are their roles?\nCan you describe what they do, and perhaps give a few examples drawn from their recent work?\n\nThe interview might proceed in different ways depending on their initial responses. Here are some questions I might ask, corresponding with the participants’ role and area of expertise.\n\n\nMaintaining the project community\n\nPlease briefly describe the process through which you obtain new partners or users.\nCan you please recall a recent example?\n\n\n\nHow well do you know each partner?\nDid you know them before their involvement?\n\n\n\nWould you describe the project as a tight knit community, or more open-ended?\n\n\n\nHow do you communicate with partners and contributors?\nWhat kinds of media or platforms do you use, and are they targeted for specific purposes? i.e. email, newsletters, social media, skype, personal communication at conferences\n\n\n\nAre there particular people in each project who you communicate with more frequently than others?\nWho are they, and why are these the people who you connect with?\n\n\n\nWhat do you consider your role or responsibility vis-a-vis the development/growth of this community?\nHow do you foster the community’s development and growth?\nDo you consider these efforts to be effective?\n\n\n\nDoes your role as someone who leads a data harmonization initiative differentiate you from other epidemiologists?\nHow has your relationship with other epidemiologists changed after initiating this project and taking on this role?\n\n\n\nReflections on data’s value\n\nHow has the data been used or analyzed?\nDo you track how the data is used?\nIs this tracking formal or informal?\n\n\n\nWhat patterns or trends are apparent based on this tracking?\nIn your view, has the data been used in productive ways?\nIn what ways are people either maximizing or not fully utilizing the data’s full potential?\n\n\n\nCan you tell me about any unexpected or alternative uses of the data?\nWhat made them significant to you?\n\n\n\nWhich skills and competencies do you think researchers need to possess in order to be able to make proper use of the data in their work?\n\n\n\nBased on your experience, what are the main obstacles for the effective and widespread adoption of these skills?\nWhat are some positive factors, or drivers, that can make that prospect more tangible?\n\n\n\nData ownership\n\nWho has rights (in the legal sense or informally) over the information contained in the system, or in related documents and datasets?\nCan you tell me about any conflicts or tensions that emerged relating to expressions of propriety or ownership over data?\n\n\n\nCollecting data\n\nDo projects collect data with future harmonization in mind?\nIf so, how does this affect data collection procedures, and does this play a role in subsequent decision-making?\n\n\n\nCurating data\n\nPlease describe the overall process of getting data into the system and then working with the data.\n\n\n\nPlease tell me about any unexpected or problematic cases that made working with data particularly challenging.\nWhat made these cases unique or challenging?\nHow did you resolve them or work towards a solution or viable outcome?\n\n\n\nAccessing data\n\nDo you consider the system easy to access?\nCan you identify some challenges that pose as barriers to access?\n\n\n\nWho has access to data?\nHow are decisions regarding access rights made?\nCan you tell me about any unnaceptable practices regarding accessing and sharing data?\n\n\n\nUsing data\n\nIf you engage with the data with specific questions in mind, how do these questions emerge?\nWhat role does the data play in shaping the questions and analytical approach?\n\n\n\nIs the system amenable to exploratory or serendipitous modes of discovery?\nPlease tell me about specific examples where you engaged with the data in this way.\n\n\n\nWhat features does the system have to view or export data?\nHow easy is it to view, export or visualize data the data?\nDo you use the tools that are designed to export of visualize data, or do you prefer to use your own tooling?\nWhat are the reasons behind this preference?\n\n\n\nDocumentation\n\nHow is the system documented?\nWho is responsible for creating documentation?\nCan you please tell me about a great example of documentation in your project?\n\n\n\nOverall, do you consider your project’s documentation to be helpful?\nWhy or why not?\n\n\n\nIn your opinion, does the documentation accurately reflect the true nature of the documented data or work practices?\nAre specific things more accurately documented than others?\nPlease tell me why you think some things are more accurately or less accurately documented.\n\n\n\nCan you recall any instances when documentation was updated?\nWhat prompted these updates?\n\n\n\nDo people ever get in touch to ask questions about specific aspects of the data or data curation procedures?\nWhat kinds of questions do they ask?\nWhat kinds of responses are given?\n\n\n\nRelationships with Maelstrom\n\nCan you please concisely describe the role of Maelstrom as part of your project’s overall initiative?\n\n\n\nWhat are the origins of your project’s relationship with Maelstrom?\nHow has this relationship changed over time?\n\n\n\nDoes your project present any unique challenges or require any special attention?\nIf so, please tell me about some unique cases or examples that demonstrate this unique relationship.\n\n\n\nDo you believe that Maelstrom is meeting your project’s needs and enabling it to achieve its goals?\nIn what ways is Maelstrom either satisfying or failing to meet your project’s expectations or needs?\nHow would you change the current system to better suit your project’s needs more effectively?\n\n\n\nDo you engage with Maelstrom’s other partners?\nIf so, what is the nature of these relationships?", + "text": "This website serves as a hub for my postdoctoral research at McGill University’s Covid-19 Immunity Task Force Databank.\nThe project is concerned with articulating social, collaborative and discursive aspects of epidemiological data-sharing initiatives, and how they relate to, overlap with or conflict with technical, institutional and epistemic factors.\nThis website hosts a series of preparatory protocols that structure the project, as well as notes about key concepts and reflections on the progress of work. Please keep in mind that this is a continually evolving site and its contents may change as the project goes on. All content is hosted and tracked at github.com/zackbatist/CITF-Postdoc.\nHere’s an overview of what’s on this site:\nContext: My motivations for doing this work and the circumstances that surround the establishment of the project.\nResearch Protocol: Outlines the project’s overall vision and contextualizes it in relation to specific objectives.\nCase Selection: Articulates the parameters that inform how cases are selected.\nEthics Protocol: Specifies ethical considerations, including risks of harm and strategies for mitigating them.\nInterview Protocol: The questions I will be asking research participants, including the rationale for asking them.\nData Management: Procedures that circumscribe collection, management and curation of research data.\nQDA Protocol: The code system, memoing guidelines, and specific QDA procedures.\nGlossary: A series of key terms and their definitions, with reference to the literature and expanded notes about their meanings.\nNotes: Some semi-structured ideas that situate my work in relation to extant literature.\nBlog: Updates and reflections on key events, or general thoughts I wish to share.\nGitHub: A link to the GitHub repository where this website’s files are hosted.\nBib: A biblatex file containing a continually-updated list of sources cited in all documents hosted on this website.\nRSS: RSS feed you can use to subscribe to the blog.", "crumbs": [ - "Interview Protocol" + "Home" ] }, { - "objectID": "glossary.html", - "href": "glossary.html", - "title": "Glossary", + "objectID": "ethics-protocol.html", + "href": "ethics-protocol.html", + "title": "Ethics Protocol", "section": "", - "text": "data-sharing\nXXXX.\n\ndata-harmonization\nXXXX.\n\ndata-integration\nXXXX.\n\ncollaboration\nXXXX.\n\ndata-sharing initiative\nXXXX.\nCorresponds with the term “harmonization initiative” in Fortier et al. (2017).\ncatalogue\nXXXX.\nBergeron et al. (2018), Bergeron et al. (2021)\n\n\n\n\nReferences\n\nBergeron, Julie, Dany Doiron, Yannick Marcon, Vincent Ferretti, and Isabel Fortier. 2018. “Fostering Population-Based Cohort Data Discovery: The Maelstrom Research Cataloguing Toolkit.” PLOS ONE 13 (7): e0200926. https://doi.org/10.1371/journal.pone.0200926.\n\n\nBergeron, Julie, Rachel Massicotte, Stephanie Atkinson, Alan Bocking, William Fraser, Isabel Fortier, and the ReACH member cohorts’ principal investigators. 2021. “Cohort Profile: Research Advancement Through Cohort Cataloguing and Harmonization (ReACH).” International Journal of Epidemiology 50 (2): 396–97. https://doi.org/10.1093/ije/dyaa207.\n\n\nFortier, Isabel, Parminder Raina, Edwin R Van den Heuvel, Lauren E Griffith, Camille Craig, Matilda Saliba, Dany Doiron, et al. 2017. “Maelstrom Research Guidelines for Rigorous Retrospective Data Harmonization.” International Journal of Epidemiology 46 (1): 103–5. https://doi.org/10.1093/ije/dyw075.", + "text": "Project Title: Articulating epidemiological data harmonization initiatives as practical and collaborative experiences\nSubmitted Materials: zackbatist.info/CITF-Postdoc/irb-docs.pdf\nPrincipal Investigator: Zachary Batist\nProtocol Number: 25-01-057\nSubmitted: 2025-01-30\nApproved:", "crumbs": [ - "Glossary" + "Ethics Protocol" ] }, { - "objectID": "data-management.html", - "href": "data-management.html", - "title": "Data Management Plan", - "section": "", - "text": "Note\n\n\n\nThis document is still a work in progress.\nThis document describes the data that this research will produce and the procedures for curating data throughout the project.", + "objectID": "ethics-protocol.html#recruitment-and-consent", + "href": "ethics-protocol.html#recruitment-and-consent", + "title": "Ethics Protocol", + "section": "Recruitment and consent", + "text": "Recruitment and consent\nWill this study involve recruitment of human study participants?\n\nYes\nNo\n\nHow are potential study participants identified and/or recruited to the study? Explain how potential participants are identified or introduced to the study, and who will recruit participants. Will the investigator/s require any special permissions or access to the target population e.g. clinic access, patient registries or records, mailing lists, community access?\nThrough consultation with key community stakeholders, the principal investigator will devise a list of prospective projects to serve as cases.1 The principal investigator will then write to the leaders of these projects inviting them to participate in the study. These invitations to project leaders will explain the project’s purpose and scope, and will encourage the recipient to reply with any questions or concerns they may have. If they accept the invitation, the principal investigator will then work with project leaders to devise a list of individuals who may serve as interview candidates based on their roles in the project. The principal investigator will be clear with project leaders that they should not pressure those who work for them to participate in the study, and that individuals’ participation should be treated as separate from their regular duties; if project leaders cannot or will not abide by this condition, their project will be rejected as a prospective case. The principal investigator will then write to the recommended individuals to introduce the study and its objectives and to invite them to participate as research subjects. If these individuals express interest in participating in the study, the principal investigator will schedule a time to sit for an interview. Some interviews may be conducted remotely using internet-based video conferencing software, depending on participants’ availability.\n1 See the case selection protocol for further details.Describe the consent process. If alternate processes for seeking consent are planned (e.g. verbal, online, waiver), please provide a rationale and outline the procedure of obtaining and documenting consent and/or assent, where applicable.\nOnce individuals express their interest in participating, participants will provided with an informed consent document that outlines in more detail the goals of the study, the roles of the participant, how they will be recorded, how data pertaining to them will be retained, and the potential risks and benefits pertaining to their involvement. This document will also describe how participants’ personally identifiable information will be managed and used. Participants will be asked to read and sign the document in order to obtain written informed consent. For interviews that will be held remotely using internet-based video conferencing software, participants will asked to send their signed informed consent documents in PDF format to the principal investigator. At the start of each interview the researcher will reiterate participants’ rights and ask them to orally reaffirm their consent before proceeding.\nIs there a relationship between the study participants and the person obtaining consent and/or the principal investigator/s?\n\nYes\nNo\n\nIf yes, please explain the nature of the relationship, and outline the steps that will be taken to avoid the perception of undue influence.\nOne project that serves as a case in this research is the Covid-19 Immunity Task Force (CITF), which the principal investigator currently serve as postdoctoral researcher. Some of the participants will therefore be his colleagues. The interviews will remain structured and limited in scope, and will not touch on matters relating to other aspects of their work. Moreover, prior to and throughout their involvement as research participants, frank and open discussion will be encouraged regarding collective expectations and to articulate the boundaries between participants’ relationships with the principal investigator as colleagues and as research subjects.\nThe principal investigator will consult with David Buckeridge, who leads the CITF, as one key community stakeholder to help devise a shortlist of projects that may serve as prospective cases.", "crumbs": [ - "Data Management" + "Ethics Protocol" ] }, { - "objectID": "data-management.html#data-collection", - "href": "data-management.html#data-collection", - "title": "Data Management Plan", - "section": "Data collection", - "text": "Data collection\nInterviews generate audio, video and textual records. I primarily rely on a SONY ICD-UX560 audio recorder, which contains 4GB of internal storage supplemented with a 32GB microSD card.1 I record using the lossless 16bit 44.1 kHz Linear PCM wav format.\n1 See https://weloty.com/sony-icd-ux560-review and https://weloty.com/using-the-sony-icd-ux560-the-4-how-tos for more information on this audio recording device.With participants’ consent, I also record video using a GoPro Hero 4 Silver action camera equipped with a 64GB microSD card.\nI maintain typed notes before, during and after each interview, which I may edit and re-organize to enhance clarity.\nImmediately after each interview, I copy the data off the recording devices and onto a dedicated project drive. I organize and rename files using a semantic naming scheme and then mirror all files onto physical and a cloud-based backup drives.", + "objectID": "ethics-protocol.html#risk-benefit-assessment", + "href": "ethics-protocol.html#risk-benefit-assessment", + "title": "Ethics Protocol", + "section": "Risk-benefit assessment", + "text": "Risk-benefit assessment\nDescribe the foreseeable risks to study participants. What risks are attributable to the research, including cumulative risks? Which risks are participants normally exposed to in the course of their clinical care or in their daily activities as they relate to the research questions/objectives?\nParticipation in this study does not involve any physical, psychological or legal risks. However, the principal investigator will be asking participants to share detailed information about their work practices and work relationships, and public association with their responses may potentially disrupt or complicate their professional reputations. To mitigate against this potential harm, the principal investigator will give participants the option to render their responses confidential.\nWhat procedures are in place to monitor and assess participant safety for the duration of the study?\nPrior to each interview, and as part of the procedure for obtaining informed consent, participants will be asked about whether they want to render their responses confidential. Immediately after each interview, participants will be given an additional opportunity to reflect on their responses, and will be prompted to either confirm or alter their decision regarding whether or not to maintain confidentiality. Furthermore, for participants who have not requested that their responses be treated as confidential immediately before and after the interview, a follow-up email will be sent one week after the interview to reiterate the option to render their responses confidential.\nDescribe the potential benefits of the study for: (1) the study participants; (2) the population under investigation, and (3) the field of research.\nThis study contributes to the development of better epidemiological data-sharing infrastructures by articulating social, collaborative and discursive aspects of data harmonization, and how these factors relate to, overlap with or conflict with technical, institutional and epistemic factors. By explicitly framing data harmonization as a social and collaborative activity, we may devise more effective data-sharing infrastructures that better support the contextualization of data and enhance their value in contexts of data reuse. This work therefore poses new ways to document how epidemiologists mobilize distributed records in the constitution of synthetic knowledge and helps develop practical solutions that enable greater reflexivity. Additionally, this study may directly benefit participants by framing the experiences they address during interviews in ways that they might not have otherwise considered, thereby encouraging greater reflexivity in their own work.", "crumbs": [ - "Data Management" + "Ethics Protocol" ] }, { - "objectID": "data-management.html#processing-data", - "href": "data-management.html#processing-data", - "title": "Data Management Plan", - "section": "Processing data", - "text": "Processing data\nI use FFmpeg and Audacity to cut, concatenate, and clean the audio and video files if necessary. I generate transcripts of audio recordings following the transcription protocol.", + "objectID": "ethics-protocol.html#privacy-and-confidentiality", + "href": "ethics-protocol.html#privacy-and-confidentiality", + "title": "Ethics Protocol", + "section": "Privacy and confidentiality", + "text": "Privacy and confidentiality\nPlease describe the measures in place for meeting confidentiality obligations. How is information and data safeguarded for the full cycle of the study: i.e. during its collection, use, dissemination, retention, and/or disposal?\nThe specific circumstances that frame each case are significant factors that will shape the findings, and the study will benefit from participants’ consent to associate their identities with their interview responses. However, they may choose to render their interview responses confidential while maintaining their role a research participant. Participants may change their decision regarding whether or not to associate their identities with their interview responses up to one week after the interview, at which point the principal investigator will begin transcribing and analyzing the records pertaining to the interview. Participants will be reminded about this option immediately after the interview and one week following the interview via email.\nThe study engages with a relatively small community, and there is minimal social risk that others may be able to determine the identities of those whose research practices and professional relationships are being documented, even if their responses are rendered confidential. To address this issue, if any single participant from a case decides to render their responses confidential, the responses of all participants pertaining to that case will be rendered confidential as well, and the identify of the project that serves as the case will be obfuscated too.\nIn situations whereby a participant decides to render their responses confidential, or has their responses rendered confidential due to another member of their case deciding to do so, only the principal investigator will have access to records containing un-obfuscated information that may identify them. These un-obfuscated records, which may include audio and video records of interview sessions, as well as unedited transcripts and textual notes containing information that may reveal the participants’ identities, will be kept in secure and encrypted media, and destroyed within five years of concluding the study, which provides sufficient time to revisit the data and produce additional research outputs. However, edited transcripts scrubbed of all information that may identify research participants may be kept, published and archived. If participants consent to maintaining association between their responses and their identities, un-obfuscated records and transcripts may be kept, published and archived.\nThe study is committed to adhering to fundamental data security practices, including those specified in McGill University’s Cloud Directive which regulates the curation of sensitive research data. Physical records will be kept in a locked drawer in secure workspaces, either at McGill University’s School of Public and Global Health or at the principal researcher’s home office. Digital records will be stored on encrypted and password-protected drives and on secure servers approved or managed by McGill University under the Cloud Directive.2\n2 Refer to the data management plan for further details on how information pertaining to this project will be collected, curated and shared.Recordings of remote interviews conducted using internet-based video conferencing software will be made using the software’s built-in recording tools. Only video conferencing software approved by the Cloud Directive will be used. Participants will be instructed to disable their microphones or video cameras prior to initiating recording if they have opted to not be recorded through these media. The researcher will record all media locally and refrain from using any cloud services to store or modify the records which the video conference software may provide.\nIf a contracted cloud/storage service provider or online survey tool is used, provide information on the service provider’s security and privacy policy, location of its servers, data ownership, and what happens to the stored data after the contract is terminated. For more information, please consult the University’s directive.\nThe study uses file-sharing software hosted by the Covid-19 Immunity Task Force at McGill University’s School of Public and Global Health to backup all files maintained for this study. These backups will include files containing information that might reveal participants’ identities. The software used to manage these backups is managed by McGill University and has been approved for storing sensitive research data by the Cloud Directive.\nThe study may use the secure GitLab instance hosted by the surveillance lab within the Clinical and Health Informatics Research Group at McGill University to store and track changes to sensitive research data. This software is managed by McGill University and has been approved for storing sensitive research data by the Cloud Directive.\nThe study maintains a website where the principal investigator shares documentation that supports the study and reflects on the work as it progresses. This is hosted using GitHub Pages and is backed up using Dropbox. No sensitive research data will pass through these services.\nRecordings of remote interviews conducted using internet-based video conferencing software will be made using the software’s built-in recording tools. Only video confering software approved by the Cloud Directive will be used. Participants will be instructed to disable their microphones or video cameras prior to initiating recording if they have opted to not be recorded through these media. The researcher will record all media locally and refrain from using any cloud services to store or modify the records which the video conference software may provide.\nPlease explain any reasonable and foreseeable disclosure requirements (e.g. disclosure to third parties such as government agencies or departments, community partners in research, personnel from an agency that monitors research, research sponsor, the REB/IRB, or regulatory agencies).\nNo disclosure requirements are foreseen.\nIf there are plans for retaining participant and/or study data for future use, please describe the context for its use, requirements for potentially re-contacting study participants and consent, and how the data will be stored and maintained for the long term.\nResearch data will be published in compliance with ethical standards for sharing open social science research data. Records that contain personally-identifying information pertaining to participants who have requested that their responses be rendered confidential and to those who have had their responses rendered confidential due to another member of their case deciding to do so will not be published.\nThe database containing codings, memos and trends deriving from qualitative data analysis will be published only after being scrubbed of all personally-identifying information pertaining to participants who have requested that their responses be rendered confidential and to those who have had their responses rendered confidential due to another member of their case deciding to do so.\nThe principal investigator may follow up with the leaders of the data-sharing initiatives that serve as cases for this project to share the results with them and to present them with constructive feedback deriving from the study’s findings. The principal investigator may also invite select participants to collaborate on a position paper advocating for reforms based on the project’s findings.\nSecondary use of data studies: if the study involves data linkage, please describe the data that will be linked and the likelihood that identifiable information will be created through the linkage.\nThis project does not rely on data deriving from other studies. The data may be reused in related work being undertaken under the same grant and by those who access the openly accessible data after they are published.", "crumbs": [ - "Data Management" + "Ethics Protocol" ] }, { - "objectID": "data-management.html#storage-and-backups", - "href": "data-management.html#storage-and-backups", - "title": "Data Management Plan", - "section": "Storage and backups", - "text": "Storage and backups\nI keep all data on a dedicated portable solid state drive which serves as a working directory for all research activities. I mirror the contents of this drive onto an identical secondary backup drive and onto cloud storage administered by the CITF using rsync.\nI maintain this website where I share documentation that supports this project and reflect on the work as it progresses. It is hosted using GitHub Pages and is backed up using Dropbox, however no sensitive research data will pass through these services.", + "objectID": "ethics-protocol.html#managing-conflicts-of-interest", + "href": "ethics-protocol.html#managing-conflicts-of-interest", + "title": "Ethics Protocol", + "section": "Managing conflicts of interest", + "text": "Managing conflicts of interest\nConflicts of interest do not imply wrong-doing. It is the responsibility of the investigator to determine if any conflicts apply to any person/s involved in the design and/or conduct of the research study or any member of their immediate family. Disclose all contracts and any conflicts of interest (real, perceived, or potential) relating to this research project. Conflict of interest may also arise with regard to the disclosure of personal health information.\n\nNot applicable. There are no conflicts of interest to disclose.\nYes, there are conflicts of interest to disclose.\n\nIf yes, please describe the conflicts of interest (real, potential, and perceived), and the procedures for managing declared conflicts. Not applicable.", "crumbs": [ - "Data Management" + "Ethics Protocol" ] }, { - "objectID": "data-management.html#publishing-and-archiving", - "href": "data-management.html#publishing-and-archiving", - "title": "Data Management Plan", - "section": "Publishing and archiving", - "text": "Publishing and archiving\nAfter my research is complete, I will make a concerted effort to document all the data I will have collected and generated and deposit them in a digital repository operated by a dedicated team of digital archivists certified to curate and preserve research data in perpetuity.", + "objectID": "context.html", + "href": "context.html", + "title": "Context", + "section": "", + "text": "I am Zack Batist — a postdoctoral researcher at McGill University, in the School of Global and Public Health’s Department of Epidemiology, Biostatistics and Occupuational Health. I’m working with David Buckeridge, who leads the Covid-19 Immunity Task Force (CITF) Databank, to investigate data sharing in epidemiological research — with an emphasis on the practical and situated experiences involved in data sharing.\nThe CITF is a “data harmonization” initiative, which entails coordinating a systematic effort to align the information contained in datasets collected by distributed teams of epidemiologists. These efforts to integrate the records collected during various discrete studies are motivated by a desire to establish larger integrated datasets bearing greater statistical power and that facilitate comparison across cohorts. However, epidemiologists must reckon with the diversity of minor variations in data collection procedures, as well as ethico-legal concerns relating to the sharing of individual health records pertaining to human research subjects across numerous institutional and regional jurisdictions.\nAs a scholar of scientific practice, with a primary interest in data-sharing and the formation of information commons, data harmonization represents a fascinating mechanism through which scientists derive technical, administrative, social and epistemic frameworks to enhance the value of their collective endeavours in response to disciplinary needs, warrants, desires and expectations. This study therefore articulates the motivations for doing data harmonization, identifies how value is ascertained, and describes the strategies employed to achieve the desired goals — including perceived and actual challenges, setbacks, opportunities, realizations, and lessons learned.\nThis relates to my previous work that (a) explores tensions that arise when attempting to establish information commons in archaeology, specifically relating to inability to cope with a superficial perception of data’s stability and an intuitive understanding of their situated nature; and that (b) investigates how the open science movement attempts (and fails) to reshape practices relating to data sharing, integration and reuse. I continue in my approach that frames data-sharing — whether it occurs in relatively “closed” curcumstances between close colleagues, or as mediated by open data platforms among strangers — as comprising a series of collaborative commitments that govern who may contribute to and obtain value from the information commons, and in what ways.", "crumbs": [ - "Data Management" + "Context" ] }, { @@ -385,227 +357,269 @@ ] }, { - "objectID": "context.html", - "href": "context.html", - "title": "Context", + "objectID": "data-management.html", + "href": "data-management.html", + "title": "Data Management Plan", "section": "", - "text": "I am Zack Batist — a postdoctoral researcher at McGill University, in the School of Global and Public Health’s Department of Epidemiology, Biostatistics and Occupuational Health. I’m working with David Buckeridge, who leads the Covid-19 Immunity Task Force (CITF) Databank, to investigate data sharing in epidemiological research — with an emphasis on the practical and situated experiences involved in data sharing.\nThe CITF is a “data harmonization” initiative, which entails coordinating a systematic effort to align the information contained in datasets collected by distributed teams of epidemiologists. These efforts to integrate the records collected during various discrete studies are motivated by a desire to establish larger integrated datasets bearing greater statistical power and that facilitate comparison across cohorts. However, epidemiologists must reckon with the diversity of minor variations in data collection procedures, as well as ethico-legal concerns relating to the sharing of individual health records pertaining to human research subjects across numerous institutional and regional jurisdictions.\nAs a scholar of scientific practice, with a primary interest in data-sharing and the formation of information commons, data harmonization represents a fascinating mechanism through which scientists derive technical, administrative, social and epistemic frameworks to enhance the value of their collective endeavours in response to disciplinary needs, warrants, desires and expectations. This study therefore articulates the motivations for doing data harmonization, identifies how value is ascertained, and describes the strategies employed to achieve the desired goals — including perceived and actual challenges, setbacks, opportunities, realizations, and lessons learned.\nThis relates to my previous work that (a) explores tensions that arise when attempting to establish information commons in archaeology, specifically relating to inability to cope with a superficial perception of data’s stability and an intuitive understanding of their situated nature; and that (b) investigates how the open science movement attempts (and fails) to reshape practices relating to data sharing, integration and reuse. I continue in my approach that frames data-sharing — whether it occurs in relatively “closed” curcumstances between close colleagues, or as mediated by open data platforms among strangers — as comprising a series of collaborative commitments that govern who may contribute to and obtain value from the information commons, and in what ways.", + "text": "Note\n\n\n\nThis document is still a work in progress.\nThis document describes the data that this research will produce and the procedures for curating data throughout the project.", "crumbs": [ - "Context" + "Data Management" ] }, { - "objectID": "ethics-protocol.html", - "href": "ethics-protocol.html", - "title": "Ethics Protocol", - "section": "", - "text": "Project Title: Articulating epidemiological data harmonization initiatives as practical and collaborative experiences\nSubmitted Materials: zackbatist.info/CITF-Postdoc/irb-docs.pdf\nPrincipal Investigator: Zachary Batist\nProtocol Number: 25-01-057\nSubmitted: 2025-01-30\nApproved:", + "objectID": "data-management.html#data-collection", + "href": "data-management.html#data-collection", + "title": "Data Management Plan", + "section": "Data collection", + "text": "Data collection\nInterviews generate audio, video and textual records. I primarily rely on a SONY ICD-UX560 audio recorder, which contains 4GB of internal storage supplemented with a 32GB microSD card.1 I record using the lossless 16bit 44.1 kHz Linear PCM wav format.\n1 See https://weloty.com/sony-icd-ux560-review and https://weloty.com/using-the-sony-icd-ux560-the-4-how-tos for more information on this audio recording device.With participants’ consent, I also record video using a GoPro Hero 4 Silver action camera equipped with a 64GB microSD card.\nI maintain typed notes before, during and after each interview, which I may edit and re-organize to enhance clarity.\nImmediately after each interview, I copy the data off the recording devices and onto a dedicated project drive. I organize and rename files using a semantic naming scheme and then mirror all files onto physical and a cloud-based backup drives.", "crumbs": [ - "Ethics Protocol" + "Data Management" ] }, { - "objectID": "ethics-protocol.html#recruitment-and-consent", - "href": "ethics-protocol.html#recruitment-and-consent", - "title": "Ethics Protocol", - "section": "Recruitment and consent", - "text": "Recruitment and consent\nWill this study involve recruitment of human study participants?\n\nYes\nNo\n\nHow are potential study participants identified and/or recruited to the study? Explain how potential participants are identified or introduced to the study, and who will recruit participants. Will the investigator/s require any special permissions or access to the target population e.g. clinic access, patient registries or records, mailing lists, community access?\nThrough consultation with key community stakeholders, the principal investigator will devise a list of prospective projects to serve as cases.1 The principal investigator will then write to the leaders of these projects inviting them to participate in the study. These invitations to project leaders will explain the project’s purpose and scope, and will encourage the recipient to reply with any questions or concerns they may have. If they accept the invitation, the principal investigator will then work with project leaders to devise a list of individuals who may serve as interview candidates based on their roles in the project. The principal investigator will be clear with project leaders that they should not pressure those who work for them to participate in the study, and that individuals’ participation should be treated as separate from their regular duties; if project leaders cannot or will not abide by this condition, their project will be rejected as a prospective case. The principal investigator will then write to the recommended individuals to introduce the study and its objectives and to invite them to participate as research subjects. If these individuals express interest in participating in the study, the principal investigator will schedule a time to sit for an interview. Some interviews may be conducted remotely using internet-based video conferencing software, depending on participants’ availability.\n1 See the case selection protocol for further details.Describe the consent process. If alternate processes for seeking consent are planned (e.g. verbal, online, waiver), please provide a rationale and outline the procedure of obtaining and documenting consent and/or assent, where applicable.\nOnce individuals express their interest in participating, participants will provided with an informed consent document that outlines in more detail the goals of the study, the roles of the participant, how they will be recorded, how data pertaining to them will be retained, and the potential risks and benefits pertaining to their involvement. This document will also describe how participants’ personally identifiable information will be managed and used. Participants will be asked to read and sign the document in order to obtain written informed consent. For interviews that will be held remotely using internet-based video conferencing software, participants will asked to send their signed informed consent documents in PDF format to the principal investigator. At the start of each interview the researcher will reiterate participants’ rights and ask them to orally reaffirm their consent before proceeding.\nIs there a relationship between the study participants and the person obtaining consent and/or the principal investigator/s?\n\nYes\nNo\n\nIf yes, please explain the nature of the relationship, and outline the steps that will be taken to avoid the perception of undue influence.\nOne project that serves as a case in this research is the Covid-19 Immunity Task Force (CITF), which the principal investigator currently serve as postdoctoral researcher. Some of the participants will therefore be his colleagues. The interviews will remain structured and limited in scope, and will not touch on matters relating to other aspects of their work. Moreover, prior to and throughout their involvement as research participants, frank and open discussion will be encouraged regarding collective expectations and to articulate the boundaries between participants’ relationships with the principal investigator as colleagues and as research subjects.\nThe principal investigator will consult with David Buckeridge, who leads the CITF, as one key community stakeholder to help devise a shortlist of projects that may serve as prospective cases.", + "objectID": "data-management.html#processing-data", + "href": "data-management.html#processing-data", + "title": "Data Management Plan", + "section": "Processing data", + "text": "Processing data\nI use FFmpeg and Audacity to cut, concatenate, and clean the audio and video files if necessary. I generate transcripts of audio recordings following the transcription protocol.", "crumbs": [ - "Ethics Protocol" + "Data Management" ] }, { - "objectID": "ethics-protocol.html#risk-benefit-assessment", - "href": "ethics-protocol.html#risk-benefit-assessment", - "title": "Ethics Protocol", - "section": "Risk-benefit assessment", - "text": "Risk-benefit assessment\nDescribe the foreseeable risks to study participants. What risks are attributable to the research, including cumulative risks? Which risks are participants normally exposed to in the course of their clinical care or in their daily activities as they relate to the research questions/objectives?\nParticipation in this study does not involve any physical, psychological or legal risks. However, the principal investigator will be asking participants to share detailed information about their work practices and work relationships, and public association with their responses may potentially disrupt or complicate their professional reputations. To mitigate against this potential harm, the principal investigator will give participants the option to render their responses confidential.\nWhat procedures are in place to monitor and assess participant safety for the duration of the study?\nPrior to each interview, and as part of the procedure for obtaining informed consent, participants will be asked about whether they want to render their responses confidential. Immediately after each interview, participants will be given an additional opportunity to reflect on their responses, and will be prompted to either confirm or alter their decision regarding whether or not to maintain confidentiality. Furthermore, for participants who have not requested that their responses be treated as confidential immediately before and after the interview, a follow-up email will be sent one week after the interview to reiterate the option to render their responses confidential.\nDescribe the potential benefits of the study for: (1) the study participants; (2) the population under investigation, and (3) the field of research.\nThis study contributes to the development of better epidemiological data-sharing infrastructures by articulating social, collaborative and discursive aspects of data harmonization, and how these factors relate to, overlap with or conflict with technical, institutional and epistemic factors. By explicitly framing data harmonization as a social and collaborative activity, we may devise more effective data-sharing infrastructures that better support the contextualization of data and enhance their value in contexts of data reuse. This work therefore poses new ways to document how epidemiologists mobilize distributed records in the constitution of synthetic knowledge and helps develop practical solutions that enable greater reflexivity. Additionally, this study may directly benefit participants by framing the experiences they address during interviews in ways that they might not have otherwise considered, thereby encouraging greater reflexivity in their own work.", + "objectID": "data-management.html#storage-and-backups", + "href": "data-management.html#storage-and-backups", + "title": "Data Management Plan", + "section": "Storage and backups", + "text": "Storage and backups\nI keep all data on a dedicated portable solid state drive which serves as a working directory for all research activities. I mirror the contents of this drive onto an identical secondary backup drive and onto cloud storage administered by the CITF using rsync.\nI maintain this website where I share documentation that supports this project and reflect on the work as it progresses. It is hosted using GitHub Pages and is backed up using Dropbox, however no sensitive research data will pass through these services.", "crumbs": [ - "Ethics Protocol" + "Data Management" ] }, { - "objectID": "ethics-protocol.html#privacy-and-confidentiality", - "href": "ethics-protocol.html#privacy-and-confidentiality", - "title": "Ethics Protocol", - "section": "Privacy and confidentiality", - "text": "Privacy and confidentiality\nPlease describe the measures in place for meeting confidentiality obligations. How is information and data safeguarded for the full cycle of the study: i.e. during its collection, use, dissemination, retention, and/or disposal?\nThe specific circumstances that frame each case are significant factors that will shape the findings, and the study will benefit from participants’ consent to associate their identities with their interview responses. However, they may choose to render their interview responses confidential while maintaining their role a research participant. Participants may change their decision regarding whether or not to associate their identities with their interview responses up to one week after the interview, at which point the principal investigator will begin transcribing and analyzing the records pertaining to the interview. Participants will be reminded about this option immediately after the interview and one week following the interview via email.\nThe study engages with a relatively small community, and there is minimal social risk that others may be able to determine the identities of those whose research practices and professional relationships are being documented, even if their responses are rendered confidential. To address this issue, if any single participant from a case decides to render their responses confidential, the responses of all participants pertaining to that case will be rendered confidential as well, and the identify of the project that serves as the case will be obfuscated too.\nIn situations whereby a participant decides to render their responses confidential, or has their responses rendered confidential due to another member of their case deciding to do so, only the principal investigator will have access to records containing un-obfuscated information that may identify them. These un-obfuscated records, which may include audio and video records of interview sessions, as well as unedited transcripts and textual notes containing information that may reveal the participants’ identities, will be kept in secure and encrypted media, and destroyed within five years of concluding the study, which provides sufficient time to revisit the data and produce additional research outputs. However, edited transcripts scrubbed of all information that may identify research participants may be kept, published and archived. If participants consent to maintaining association between their responses and their identities, un-obfuscated records and transcripts may be kept, published and archived.\nThe study is committed to adhering to fundamental data security practices, including those specified in McGill University’s Cloud Directive which regulates the curation of sensitive research data. Physical records will be kept in a locked drawer in secure workspaces, either at McGill University’s School of Public and Global Health or at the principal researcher’s home office. Digital records will be stored on encrypted and password-protected drives and on secure servers approved or managed by McGill University under the Cloud Directive.2\n2 Refer to the data management plan for further details on how information pertaining to this project will be collected, curated and shared.Recordings of remote interviews conducted using internet-based video conferencing software will be made using the software’s built-in recording tools. Only video conferencing software approved by the Cloud Directive will be used. Participants will be instructed to disable their microphones or video cameras prior to initiating recording if they have opted to not be recorded through these media. The researcher will record all media locally and refrain from using any cloud services to store or modify the records which the video conference software may provide.\nIf a contracted cloud/storage service provider or online survey tool is used, provide information on the service provider’s security and privacy policy, location of its servers, data ownership, and what happens to the stored data after the contract is terminated. For more information, please consult the University’s directive.\nThe study uses file-sharing software hosted by the Covid-19 Immunity Task Force at McGill University’s School of Public and Global Health to backup all files maintained for this study. These backups will include files containing information that might reveal participants’ identities. The software used to manage these backups is managed by McGill University and has been approved for storing sensitive research data by the Cloud Directive.\nThe study may use the secure GitLab instance hosted by the surveillance lab within the Clinical and Health Informatics Research Group at McGill University to store and track changes to sensitive research data. This software is managed by McGill University and has been approved for storing sensitive research data by the Cloud Directive.\nThe study maintains a website where the principal investigator shares documentation that supports the study and reflects on the work as it progresses. This is hosted using GitHub Pages and is backed up using Dropbox. No sensitive research data will pass through these services.\nRecordings of remote interviews conducted using internet-based video conferencing software will be made using the software’s built-in recording tools. Only video confering software approved by the Cloud Directive will be used. Participants will be instructed to disable their microphones or video cameras prior to initiating recording if they have opted to not be recorded through these media. The researcher will record all media locally and refrain from using any cloud services to store or modify the records which the video conference software may provide.\nPlease explain any reasonable and foreseeable disclosure requirements (e.g. disclosure to third parties such as government agencies or departments, community partners in research, personnel from an agency that monitors research, research sponsor, the REB/IRB, or regulatory agencies).\nNo disclosure requirements are foreseen.\nIf there are plans for retaining participant and/or study data for future use, please describe the context for its use, requirements for potentially re-contacting study participants and consent, and how the data will be stored and maintained for the long term.\nResearch data will be published in compliance with ethical standards for sharing open social science research data. Records that contain personally-identifying information pertaining to participants who have requested that their responses be rendered confidential and to those who have had their responses rendered confidential due to another member of their case deciding to do so will not be published.\nThe database containing codings, memos and trends deriving from qualitative data analysis will be published only after being scrubbed of all personally-identifying information pertaining to participants who have requested that their responses be rendered confidential and to those who have had their responses rendered confidential due to another member of their case deciding to do so.\nThe principal investigator may follow up with the leaders of the data-sharing initiatives that serve as cases for this project to share the results with them and to present them with constructive feedback deriving from the study’s findings. The principal investigator may also invite select participants to collaborate on a position paper advocating for reforms based on the project’s findings.\nSecondary use of data studies: if the study involves data linkage, please describe the data that will be linked and the likelihood that identifiable information will be created through the linkage.\nThis project does not rely on data deriving from other studies. The data may be reused in related work being undertaken under the same grant and by those who access the openly accessible data after they are published.", + "objectID": "data-management.html#publishing-and-archiving", + "href": "data-management.html#publishing-and-archiving", + "title": "Data Management Plan", + "section": "Publishing and archiving", + "text": "Publishing and archiving\nAfter my research is complete, I will make a concerted effort to document all the data I will have collected and generated and deposit them in a digital repository operated by a dedicated team of digital archivists certified to curate and preserve research data in perpetuity.", "crumbs": [ - "Ethics Protocol" + "Data Management" ] }, { - "objectID": "ethics-protocol.html#managing-conflicts-of-interest", - "href": "ethics-protocol.html#managing-conflicts-of-interest", - "title": "Ethics Protocol", - "section": "Managing conflicts of interest", - "text": "Managing conflicts of interest\nConflicts of interest do not imply wrong-doing. It is the responsibility of the investigator to determine if any conflicts apply to any person/s involved in the design and/or conduct of the research study or any member of their immediate family. Disclose all contracts and any conflicts of interest (real, perceived, or potential) relating to this research project. Conflict of interest may also arise with regard to the disclosure of personal health information.\n\nNot applicable. There are no conflicts of interest to disclose.\nYes, there are conflicts of interest to disclose.\n\nIf yes, please describe the conflicts of interest (real, potential, and perceived), and the procedures for managing declared conflicts. Not applicable.", + "objectID": "glossary.html", + "href": "glossary.html", + "title": "Glossary", + "section": "", + "text": "data-sharing\nXXXX.\n\ndata-harmonization\nXXXX.\n\ndata-integration\nXXXX.\n\ncollaboration\nXXXX.\n\ndata-sharing initiative\nXXXX.\nCorresponds with the term “harmonization initiative” in Fortier et al. (2017).\ncatalogue\nXXXX.\nBergeron et al. (2018), Bergeron et al. (2021)\n\n\n\n\nReferences\n\nBergeron, Julie, Dany Doiron, Yannick Marcon, Vincent Ferretti, and Isabel Fortier. 2018. “Fostering Population-Based Cohort Data Discovery: The Maelstrom Research Cataloguing Toolkit.” PLOS ONE 13 (7): e0200926. https://doi.org/10.1371/journal.pone.0200926.\n\n\nBergeron, Julie, Rachel Massicotte, Stephanie Atkinson, Alan Bocking, William Fraser, Isabel Fortier, and the ReACH member cohorts’ principal investigators. 2021. “Cohort Profile: Research Advancement Through Cohort Cataloguing and Harmonization (ReACH).” International Journal of Epidemiology 50 (2): 396–97. https://doi.org/10.1093/ije/dyaa207.\n\n\nFortier, Isabel, Parminder Raina, Edwin R Van den Heuvel, Lauren E Griffith, Camille Craig, Matilda Saliba, Dany Doiron, et al. 2017. “Maelstrom Research Guidelines for Rigorous Retrospective Data Harmonization.” International Journal of Epidemiology 46 (1): 103–5. https://doi.org/10.1093/ije/dyw075.", "crumbs": [ - "Ethics Protocol" + "Glossary" ] }, { - "objectID": "index.html", - "href": "index.html", - "title": "CITF-Postdoc", + "objectID": "interview-protocol.html", + "href": "interview-protocol.html", + "title": "Interview Protocol", "section": "", - "text": "This website serves as a hub for my postdoctoral research at McGill University’s Covid-19 Immunity Task Force Databank.\nThe project is concerned with articulating social, collaborative and discursive aspects of epidemiological data-sharing initiatives, and how they relate to, overlap with or conflict with technical, institutional and epistemic factors.\nThis website hosts a series of preparatory protocols that structure the project, as well as notes about key concepts and reflections on the progress of work. Please keep in mind that this is a continually evolving site and its contents may change as the project goes on. All content is hosted and tracked at github.com/zackbatist/CITF-Postdoc.\nHere’s an overview of what’s on this site:\nContext: My motivations for doing this work and the circumstances that surround the establishment of the project.\nResearch Protocol: Outlines the project’s overall vision and contextualizes it in relation to specific objectives.\nCase Selection: Articulates the parameters that inform how cases are selected.\nEthics Protocol: Specifies ethical considerations, including risks of harm and strategies for mitigating them.\nInterview Protocol: The questions I will be asking research participants, including the rationale for asking them.\nData Management: Procedures that circumscribe collection, management and curation of research data.\nQDA Protocol: The code system, memoing guidelines, and specific QDA procedures.\nGlossary: A series of key terms and their definitions, with reference to the literature and expanded notes about their meanings.\nNotes: Some semi-structured ideas that situate my work in relation to extant literature.\nBlog: Updates and reflections on key events, or general thoughts I wish to share.\nGitHub: A link to the GitHub repository where this website’s files are hosted.\nBib: A biblatex file containing a continually-updated list of sources cited in all documents hosted on this website.\nRSS: RSS feed you can use to subscribe to the blog.", + "text": "Interviews may be held either in-person or through online video conference. All interviews will be held in quiet and comfortable environments, such as office spaces or conference rooms.\nI will record all in-person interviews using a SONY ICD-UX560 audio recorder to capture audio in the lossless 16 bit 44.1 kHz Linear PCM wav format, with additional audio filters to enhance playback during transcription, if necessary.\nI will also record in-person interview sessions using a GoPro Hero 4 Silver action camera, depending on participants willingness to be video recorded. Based on prior interviews with scientists about their research experiences, I found that interviewees like to show me, rather than merely tell me, about what they are working on and the means through which they engage with information systems. The camera may be leveraged to record spontaneous video records of these demonstrations and provide me with an additional rich data source for further analysis. Moreover, the camera provides an additional backup audio recording in case of data loss on the primary recording device.\nRemote interviews will be recorded using the video conferencing software’s built-in recording tools. Participants will be instructed to disable their microphones or video cameras prior to initiating recording if they have opted to not be recorded through these media. The researcher will record all media locally and refrain from using any cloud services to store or modify the records which the video conference software may provide.\nI will also record handwritten notes comprising descriptive accounts of activities and interactions when recording devices are switched off, as well as preliminary interpretations of observed behaviours and notes on things I plan to follow up on at a later time.", "crumbs": [ - "Home" + "Interview Protocol" ] }, { - "objectID": "notes.html", - "href": "notes.html", - "title": "Notes", + "objectID": "interview-protocol.html#interview-records", + "href": "interview-protocol.html#interview-records", + "title": "Interview Protocol", "section": "", - "text": "Modified\n\n\nTitle\n\n\nCategories\n\n\n\n\n\n\nFeb 20, 2025\n\n\nMethodology notes\n\n\nreading, general thoughts\n\n\n\n\nFeb 4, 2025\n\n\nPotential cases\n\n\ncases, brainstorming\n\n\n\n\nJan 28, 2025\n\n\nMaelstrom reading notes\n\n\nreading\n\n\n\n\n\nNo matching items", + "text": "Interviews may be held either in-person or through online video conference. All interviews will be held in quiet and comfortable environments, such as office spaces or conference rooms.\nI will record all in-person interviews using a SONY ICD-UX560 audio recorder to capture audio in the lossless 16 bit 44.1 kHz Linear PCM wav format, with additional audio filters to enhance playback during transcription, if necessary.\nI will also record in-person interview sessions using a GoPro Hero 4 Silver action camera, depending on participants willingness to be video recorded. Based on prior interviews with scientists about their research experiences, I found that interviewees like to show me, rather than merely tell me, about what they are working on and the means through which they engage with information systems. The camera may be leveraged to record spontaneous video records of these demonstrations and provide me with an additional rich data source for further analysis. Moreover, the camera provides an additional backup audio recording in case of data loss on the primary recording device.\nRemote interviews will be recorded using the video conferencing software’s built-in recording tools. Participants will be instructed to disable their microphones or video cameras prior to initiating recording if they have opted to not be recorded through these media. The researcher will record all media locally and refrain from using any cloud services to store or modify the records which the video conference software may provide.\nI will also record handwritten notes comprising descriptive accounts of activities and interactions when recording devices are switched off, as well as preliminary interpretations of observed behaviours and notes on things I plan to follow up on at a later time.", "crumbs": [ - "Notes" + "Interview Protocol" ] }, { - "objectID": "notes/methodology-notes.html", - "href": "notes/methodology-notes.html", - "title": "Methodology notes", + "objectID": "interview-protocol.html#interview-guide", + "href": "interview-protocol.html#interview-guide", + "title": "Interview Protocol", + "section": "Interview guide", + "text": "Interview guide\nInterviews are oriented by my goal to document values and attitudes concerning data harmonization efforts, as elicited by research participants in their responses. Participants will be asked to reflect on:\n\nthe motivations for their initiatives;\nthe challenges they experience;\nhow they envision success and failure;\ntheir perceptions of their own roles and the roles of other team members and stakeholders;\nthe values that inform their decisions;\nhow the technological apparatus they set up enables them to realize their goals and values; and\nways in which they believe data-sharing could be improved.\n\nTo this end, each interview will proceed following a strategic order:\n\n1. Participants’ goals and perspectives\nFollow a life-history method to better understand participants’ professional backgrounds and their roles within their projects. The goal is to obtain information about their paths, not the rehearsed origin story.\n\nTo start, can you please tell me a little bit about your background?\nWhat is the project, and what is your role?\nHow did you find yourself in this role?\nHow has your previous experience prepared you for your role?\n\n\n\n2. Projects’ missions, purposes, motivations\nThis section is about the project in general, including its purpose, scope and value. Information about practices and procedures will be sought in a subsequent phase of the interview.\n\nWhat are the project’s goals?\nWhat makes the project unique?\nWhat is the project doing that no other similar project is doing?\nDo you consider this project as similar to any other initiatives?\nWhat are they, and in what ways are they simiar or different?\n\n\n\nWhat are the expected outcomes?\nHave you achieved these goals and outcomes?\nIf not Are you on track to achieving them?\nWhat are some challenges that the project experienced, and how have you worked to overcome them?\n\n\n\n3. Practices, procedures, relationships\nThis section asks about specific actions and interactions that the participant engages in.\n\nRoles and relationships\n\nWhat does your role entail?\nCan you provide a couple examples of things that you recently did in this capacity?\n\n\n\nWho else do you frequently rely on, and what are their roles?\nCan you describe what they do, and perhaps give a few examples drawn from their recent work?\n\nThe interview might proceed in different ways depending on their initial responses. Here are some questions I might ask, corresponding with the participants’ role and area of expertise.\n\n\nMaintaining the project community\n\nPlease briefly describe the process through which you obtain new partners or users.\nCan you please recall a recent example?\n\n\n\nHow well do you know each partner?\nDid you know them before their involvement?\n\n\n\nWould you describe the project as a tight knit community, or more open-ended?\n\n\n\nHow do you communicate with partners and contributors?\nWhat kinds of media or platforms do you use, and are they targeted for specific purposes? i.e. email, newsletters, social media, skype, personal communication at conferences\n\n\n\nAre there particular people in each project who you communicate with more frequently than others?\nWho are they, and why are these the people who you connect with?\n\n\n\nWhat do you consider your role or responsibility vis-a-vis the development/growth of this community?\nHow do you foster the community’s development and growth?\nDo you consider these efforts to be effective?\n\n\n\nDoes your role as someone who leads a data harmonization initiative differentiate you from other epidemiologists?\nHow has your relationship with other epidemiologists changed after initiating this project and taking on this role?\n\n\n\nReflections on data’s value\n\nHow has the data been used or analyzed?\nDo you track how the data is used?\nIs this tracking formal or informal?\n\n\n\nWhat patterns or trends are apparent based on this tracking?\nIn your view, has the data been used in productive ways?\nIn what ways are people either maximizing or not fully utilizing the data’s full potential?\n\n\n\nCan you tell me about any unexpected or alternative uses of the data?\nWhat made them significant to you?\n\n\n\nWhich skills and competencies do you think researchers need to possess in order to be able to make proper use of the data in their work?\n\n\n\nBased on your experience, what are the main obstacles for the effective and widespread adoption of these skills?\nWhat are some positive factors, or drivers, that can make that prospect more tangible?\n\n\n\nData ownership\n\nWho has rights (in the legal sense or informally) over the information contained in the system, or in related documents and datasets?\nCan you tell me about any conflicts or tensions that emerged relating to expressions of propriety or ownership over data?\n\n\n\nCollecting data\n\nDo projects collect data with future harmonization in mind?\nIf so, how does this affect data collection procedures, and does this play a role in subsequent decision-making?\n\n\n\nCurating data\n\nPlease describe the overall process of getting data into the system and then working with the data.\n\n\n\nPlease tell me about any unexpected or problematic cases that made working with data particularly challenging.\nWhat made these cases unique or challenging?\nHow did you resolve them or work towards a solution or viable outcome?\n\n\n\nAccessing data\n\nDo you consider the system easy to access?\nCan you identify some challenges that pose as barriers to access?\n\n\n\nWho has access to data?\nHow are decisions regarding access rights made?\nCan you tell me about any unnaceptable practices regarding accessing and sharing data?\n\n\n\nUsing data\n\nIf you engage with the data with specific questions in mind, how do these questions emerge?\nWhat role does the data play in shaping the questions and analytical approach?\n\n\n\nIs the system amenable to exploratory or serendipitous modes of discovery?\nPlease tell me about specific examples where you engaged with the data in this way.\n\n\n\nWhat features does the system have to view or export data?\nHow easy is it to view, export or visualize data the data?\nDo you use the tools that are designed to export of visualize data, or do you prefer to use your own tooling?\nWhat are the reasons behind this preference?\n\n\n\nDocumentation\n\nHow is the system documented?\nWho is responsible for creating documentation?\nCan you please tell me about a great example of documentation in your project?\n\n\n\nOverall, do you consider your project’s documentation to be helpful?\nWhy or why not?\n\n\n\nIn your opinion, does the documentation accurately reflect the true nature of the documented data or work practices?\nAre specific things more accurately documented than others?\nPlease tell me why you think some things are more accurately or less accurately documented.\n\n\n\nCan you recall any instances when documentation was updated?\nWhat prompted these updates?\n\n\n\nDo people ever get in touch to ask questions about specific aspects of the data or data curation procedures?\nWhat kinds of questions do they ask?\nWhat kinds of responses are given?\n\n\n\nRelationships with Maelstrom\n\nCan you please concisely describe the role of Maelstrom as part of your project’s overall initiative?\n\n\n\nWhat are the origins of your project’s relationship with Maelstrom?\nHow has this relationship changed over time?\n\n\n\nDoes your project present any unique challenges or require any special attention?\nIf so, please tell me about some unique cases or examples that demonstrate this unique relationship.\n\n\n\nDo you believe that Maelstrom is meeting your project’s needs and enabling it to achieve its goals?\nIn what ways is Maelstrom either satisfying or failing to meet your project’s expectations or needs?\nHow would you change the current system to better suit your project’s needs more effectively?\n\n\n\nDo you engage with Maelstrom’s other partners?\nIf so, what is the nature of these relationships?", + "crumbs": [ + "Interview Protocol" + ] + }, + { + "objectID": "notes/maelstrom-readings.html", + "href": "notes/maelstrom-readings.html", + "title": "Maelstrom reading notes", "section": "", - "text": "This document is an overview of methodological topics and concerns. It is a place where I think through and justify my methodological decisions, and identify the methods and procedures through which I implement them." + "text": "Initial overview of data harmonization procedures, using the Healthy Obesity Project (HOP) as an illustrative case.\nOutlines the technical apparatus, especially for DataShield, but also broadly describes the discursive process of arriving at a DataSchema that is both functional and flexible.\n\nThis description is quite broad and abstracy, seems somewhat ideal and aspirational.\n\nDescribes reliance on international standards, such as the International Labour Organization’s International Standard Classification of Occupations.\n\nIt seems like these are used as black boxes that encapsulate a series of tensions which epidemiologists are unconcerned with; in effect, they simplify the need for stretching the collaborative ties even further than they are already extended, they represent matters out of scope for deeper discursive engagement.\n\nIt is notable that they emphasize that it’s easy to set up and use DataShield and Maelstorm toolkits independently of university IT and that it can be run using RStudio installed on a basic laptop.\n\nMaybe look into the historical context (2013) and the evolving role of university IT in software selection.\n\nThe conclusion states that the HOP project was successful in its harmonization efforts, but does not go as far as to state that it produced meaningful findings as a result of harmonization.\n\nI may take some time to find and read studies that used these data to see what’s what.\nThis seems like the main one: https://doi.org/10.1186/1472-6823-14-9, but these other papers may or not not also be relevant:\n\nhttps://doi.org/10.1016/j.smhl.2021.100263\nhttps://doi.org/10.1007/s10654-014-9977-1\nhttps://doi.org/10.1530/EJE-14-0540\nhttps://doi.org/10.1007/S13679-020-00375-0\nhttps://doi.org/10.1093/eurpub/ckac061" }, { - "objectID": "notes/methodology-notes.html#multicase-studies", - "href": "notes/methodology-notes.html#multicase-studies", - "title": "Methodology notes", - "section": "Multicase Studies", - "text": "Multicase Studies\nThese notes describe the features, affordances and limitations of case study research, and articules factors correspoding with variable kinds of case studies.\nI do notice a distinction between two schools of thought, which seem to be spearheaded by Stake and Yin. I generally favour Stake’s flexible approach, and it seems well aligned with other methodological works I’ve been reading (e.g. Abbott 2004; Charles C. Ragin and Becker 1992).\n\nStake’s Approach\nIn case-study research, cases represent discrete instances of a phenomenon that inform the researcher about it. The cases are not the subjects of inquiry, and instead represent unique sets of circumstances that frame or contextualize the phenomenon of interest (Stake 2006: 4-7).\nCases usually share common reference to the overall research themes, but exhibit variations that enable a researcher to capture different outlooks or perspectives on matters of common concern. Drawing from multiple cases thus enables comprehensive coverage of a broad topic that no single case may cover on its own (Stake 2006: 23). In other words, cases are contexts that ascribe particular local flavours to the activities I trace, and which I must consider to account fully for the range of motivations, circumstances and affordances that back decisions to perform activities and to implement them in specific ways.\nMoreover, the power of case study research derives from identifying consistencies that relate cases to each other, while simultaneously highlighting how their unique and distinguishing facets contribute to their representativeness of the underlying phenomon. Case study research therefore plays on the tensions that challenge relationships among cases and the phenomenon that they are being called upon to represent (C. C. Ragin 1999: 1139-1140).\nStake (2006: 4-6) uses the term quintain1 to describe the group, category or phenomenon that bind together a collection of cases. A quintain is an object, phenomenon or condition to be studied – “a target, not a bull’s eye” (Stake 2006: 6). “The quintain is the arena or holding company or umbrella for the cases we will study” (Stake 2006: 6). The quintain is the starting point for multi-case research.\nAccording to Stake (2006: 6):\n\nMulticase research starts with the quintain. To understand it better, we study some of its single cases — its sites or manifestations. But it is the quintain we seek to understand. We study what is similar and different about the cases in order to understand the quintain better.\n\nStake (2006: 8) then goes on:\n\nWhen the purpose of a case is to go beyond the case, we call it an “instrumental” case study When the main and enduring interest is in the case itself, we call it “intrinsic” case study (Stake 1988). With multicase study and its strong interest in the quintain, the interest in the cases will be primarily instrumental.\n\nAbbott’s (2004: 22) characaterization of Small-N comparison is very reminiscient of Stake’s (2006) account of the case-quintain dialectic:\n\nSmall-N comparison attempts to combine the advantages of single-case analysis with those of multicase analysis, at the same time trying to avoid the disadantages of each. On the one hand, it retains much information about each case. On the other, it compares the different cases to test arguments in ways that are impossible with a single case. By making these detailed comparisons, it tries to avoid the standard critcism of single-case analysis — that one can’t generalize from a single case — as well as the standard criticism of multicase analysis — that it oversimplifies and changes the meaning of variables by removing them from their context.\n\nIt should be noted that case study research limits my ability to define causal relationships or to derive findings that may be generalized across the whole field of epidemiology. This being said, case study research allows me to articulate the series of inter-woven factors that impact how epidedemiological researchers coordinate and participate in data-sharing initiatives, while explicitly accounting for and drawing from the unique and situational contexts that frame each case.\nStake (2006: 23) recommends selecting between 4-10 cases and identifies three main criteria for selecting cases:\n\nIs the case relevant to the quintain?\nDo the cases provide diversity across contexts?\nDo the cases provide good opportunities to learn about complexity and contexts?\n\n\nFor qualitative fieldwork, we will usually draw a purposive sample of cases, a sample tailored to our study; this will build in variety and create opportunities for intensive study (Stake 2006: 24).2\n\nStake’s (2010: 122) prioritizes doing research to understand something or to improve something, and I generally agree with his rationalization; research helps reframe problems and establish different decision options.\n\n\nYin’s Approach\nAccording to Yin (2014: 16), “a case study is an empirical inquiry that investigates a contemporary phenomenon (the ‘case’) in depth and within its real-world context, especially when the boundaries between phenomenon and context may not be clearly evident.”\nHe goes on to document some features of a case study: “A case study inquiry copes with the technically distinctive situation in which there will be many more variables of interest than data points, and as one result relies on multiple sources of evidence, with data needing to converge in a triangulating fashion, and as another result benefits from the prior development of theoretical propositions to guide data collection and analysis.” (Yin 2014: 17)\nYin (2014) is more oriented toward what he refers to as a realist perspective, which he pits against relativist and interpretivist perspectives (used interchangably, it seems), and which I might refer to as constructivist. He characterizes relativist perspectives as “acknowledging multiple realities having multiple meanings, with findings that are observer dependent”. His prioriting of a realist approach corresponds with the analysis by Yazan (2015), who compared Yin with Stake and Merriam. According to Yazan (2015: 137), Yin evades making statements about his epistemic commitments, and is characterized as post-positivist.\nYin (2014) is very concerned with research design in case study research He posits that, in a colloquial sense, “a research design is a logical plan for getting from here to there, where here may be defined as the initial set of questions to be answered, and there is some set of conclusions (answers) about these questions.” (Yin 2014: 28)\nYin distinguishes between a research design and a work plan. A research design deals with a logical problem, whereas a work plan deals with a logistical problem. Seems reminiscient of Brian Cantwell Smith’s distinction between skeletons and outlines.\nYin lists five components of a research design:\n\nA case study’s questions;\nits propositions, if any;\nits unit(s) of analysis;\nthe logic linking the data to the propositions; and\nthe criteria for interpreting the findings.\n\nInterestingly, I have been instinctively following these steps, and am currently hovering somewhere between components 3 and 4, while dipping back to 2 once in a while too.\nThe problem of defining the unit of analysis is salient to me right now. According to Yin (2014: 32), the unit of analysis may change as the project progresses, depending on initial misconceptions (he uses the example of a unit of analysis changing from neighbourhoods to small groups, as contextualized by the socio-geographical entity of the neighbourhood, which is laden with issues of class, race, etc). In my own situation, the unit of analysis may hover between the harmonization initiative, the people, activities or infrastructures that make it work.\nIn the section on criteria for interpreting the findings, Yin emphasizes the role of rival theories, which is akin to a concern with falsifiability as a means of validating truth claims, and which betrays his positivist leanings. This may be compared with Stake’s emphasis on triangulation, which is more concerned with internal cohesiveness. Similarly, Yin cites Corbin and Strauss regarding the role of theory or theoretical propositions in research design, which similarly reveals a concern with rigorous upfront planning and strict adherence to research design as a key aspect of deriving valid findings.\nRegarding generalizability, Yin (2014: 40-41) states that “Rather than thinking about your case as a sample, you should think of it as the opportunity to shed empirical light about some theoretical concepts or principles, not unlike the motive of a laboratory investigator in conceiving of and then conducting a new experiment.” He goes on to state that case studies tend to strive for analytic generalizations that go beyond the specific case that has been studied, and which apply to other concrete situations rather than just abstract theory building.\n\n\nLogistics of case study design\n\nPreparing to select case study data\nYin (2014: 72-73) identifies five desired attributes for collecting case study data:\n\nAsk good questions — and interpret answers fairly.\n\n\n“As you collect case study evidence, you must quickly review the evidence and continually ask yourself why events or perceptions appear as they do.” (73)\nA good indicator of having asked good questions is mental and emotional exhaustion at the end of each fieldwork day, due to the depletion of “analytic energy” associated with being attention on your toes. (73-74)\n\n\nBe a good “listener” not trapped by existing ideologies or preconceptions.\n\n\nSensing through multiple modalities, not just spoken words.\nAlso subtext, as elicited through choices of terms used, mood and affective components. (74)\n\n\nStay adaptive, so that newly encountered situations can be seen as opportunities, not threats.\n\n\nRemember the original purpose but willing to adapt to unanticipated circumnstances. (74)\nEmphasize balancing adaptability with rigour, but not with rigidity. (75)\n\n\nHave a firm grasp of what is being studied, even when in an exploratory mode.\n\n\nNeed to do more than merely record data, but interpret information as they are being collected and to know immedately whether there are contradictions or complementary statements to follow-up on. (75-76)\n\n\nAvoid biases of being sensitive to contrary evidence, also knowing how to conduct research ethically.\n\n\nMaintain strong professional competence, including keeping up with related research, ensuring accuracy, striving for credibility, and knowledging and mitigating against bias.\n\nYin advocates for adoption of case study protocols. He provides an example of a table of contents for case study protocols, which generally comprise four sections:\n\nOverview of the case study\nData collection procedures\nData collection questions\nGuide for the case study report\n\n\n\nTriangulation\nTriangulation is a process of gaining assurance. Also sometimes called crystallization.\n“Each important finding needs to have at least three (often more) confirmations and assurances that key meanings are not being overlooked.” (Stake 2006: 33) Triangulation is a process of repetitous data gathering and critical review of what is being said. (Stake 2006: 34)\nWhat needs triangulation? (Stake 2006: 35-36)\n\nIf the description is trivial or beyond question, there is no need to triangulate.\nIf the description is relevant and debatable, there is much need to triangulate.\nIf the data are critical to a main assertion, there is much need to triangulate.\nIf the data are evidence for a controversial finding, there is much need to triangulate.\nIf a statement is clearly a speaker’s interpretation, there is little need to triangulate the quotation but not its content.\n\nStake (2006: 37) cites Denzin (1989) who highlighted several kinds of triangulation, leading to a few advisories:\n\nFind ways to use multiple rather than single observers of the same thing.\nUse second and third perspectives, i.e. the views of teachers, student and parents.\nUse more than one research method on the same thing, i.e. document review and interview.\nCheck carefully to decide how much the total description warrants generalization.\n\nDo your conclusions generalize across other times or places?\nDo your conclusions about the aggregate generalize to individuals?\nDo findings of the interaction among individuals in one group pertain to other groups?\nDo findings of the aggregate of these people generalized to a population?\n\n\n\n\nCross-Case Analysis Procedure\nStake (2006: Chapter 3) lays out a procedure for deriving synthetic findings from data collected across cases. He frames this in terms of a dialectic between cases and quintains. He identifies three tracks (Stake 2006: 46):\n\nTrack 1: Maintains the case findings and the situationality.\nTrack 2: Merges similar findings, maintaining a little of the situationality.\nTrack 3: The most quanitative track, shifts the focus from findings to factors.\n\nAccording to Stake, case reports should be created independently and then brought together by a single individual when working in a collaborative project. In keeping with the case-quintain dialectic, this integration must involve strategically putting the cases aside and bringing them back in to identify convergences and divergences, similarities and differences, normalitities and discrepancies among them.\nThere is some detailed discussion about different kinds of statements, i.e. themes, findings, factors and assertions, but I find this a bit too much detail for me to get at at this point in mymethodological planning. In general though, Stake documents a process whereby an analyst navigates back and forth between the general and the situational, presenting tentativr statements that are shored up, modified or discarded through testing compatability of the evidence across cases.\n\n\nSingle cases\nStake (2000) is concerned with identifying what can be learned from a single case. He (2000: 437) identifies three kinds of cases:\n\nIntrinsic case studies as being driven by a desire to understand the particular case.\nInstrumental case studies are examined “mainly to provide insight into an issue or to redraw a generalization.”\nCollective case studies “investigate a phenomenon, population or general condition”.\n\nStake (2000) frames case research around a tension between the particular and the general, which echoes the case-quintain dilemma he described in (Stake 2006: 4-6).\n\n\nSome scattered practical guidance\nStake (2006: 18-22) provides a detailed and realistic overview of common challenges involved in collaborative qualitative research. This could be handy in future work when planning a multicase project involving multiple researchers.\nStake (2006: 29-33) provides guidance on how to plan and conduct interviews in multicase research, including a series of helpful prompts and questions to ask yourself while designing the interview. One thing that stands out is his recommendation that an interview should be more about the interviewee than about the case. It’s necessary to find out about the interviewee to understand their interpretations, but what they reveal about the quintain is more important.\nOn page 34, Stake (2006) also provides some practical tips for documenting and storing data, after Huberman and Miles (1994).\nStake (2006: Chapter 4) includes a chapter on procedures for reporting the findings, and I may return to this later on once I need to initiative this phase of work. It addresses concerns about how to articulate comparisons, concerns about generalization, and how to handle advocacy based on findings.\nSee Stake (2006) Chapter 5 for a step-by-step overview of a multicase study analysis. The rest of the volume after that includes three very detailed examples from his own work." + "objectID": "notes/maelstrom-readings.html#doiron2013", + "href": "notes/maelstrom-readings.html#doiron2013", + "title": "Maelstrom reading notes", + "section": "", + "text": "Initial overview of data harmonization procedures, using the Healthy Obesity Project (HOP) as an illustrative case.\nOutlines the technical apparatus, especially for DataShield, but also broadly describes the discursive process of arriving at a DataSchema that is both functional and flexible.\n\nThis description is quite broad and abstracy, seems somewhat ideal and aspirational.\n\nDescribes reliance on international standards, such as the International Labour Organization’s International Standard Classification of Occupations.\n\nIt seems like these are used as black boxes that encapsulate a series of tensions which epidemiologists are unconcerned with; in effect, they simplify the need for stretching the collaborative ties even further than they are already extended, they represent matters out of scope for deeper discursive engagement.\n\nIt is notable that they emphasize that it’s easy to set up and use DataShield and Maelstorm toolkits independently of university IT and that it can be run using RStudio installed on a basic laptop.\n\nMaybe look into the historical context (2013) and the evolving role of university IT in software selection.\n\nThe conclusion states that the HOP project was successful in its harmonization efforts, but does not go as far as to state that it produced meaningful findings as a result of harmonization.\n\nI may take some time to find and read studies that used these data to see what’s what.\nThis seems like the main one: https://doi.org/10.1186/1472-6823-14-9, but these other papers may or not not also be relevant:\n\nhttps://doi.org/10.1016/j.smhl.2021.100263\nhttps://doi.org/10.1007/s10654-014-9977-1\nhttps://doi.org/10.1530/EJE-14-0540\nhttps://doi.org/10.1007/S13679-020-00375-0\nhttps://doi.org/10.1093/eurpub/ckac061" }, { - "objectID": "notes/methodology-notes.html#grounded-theory", - "href": "notes/methodology-notes.html#grounded-theory", - "title": "Methodology notes", - "section": "Grounded theory", - "text": "Grounded theory\nThese notes are largely drawn from Charmaz (2000), which I understand to be a fairly balanced and comprehensive overview of the Glaser / Strauss and Corbin debate, and of the situation of specific methods and techniques in relation to these different stances. I also value Charmaz’s position as someone who subscribes to her constructivist approach.\nAccording to Charmaz(2000: 509):\n\nEssentially, grounded theory methods consist of systematic inductive guidelines for collecting and analyzing data to build middle-range theoretical frameworks that explain the collected data.\n\nCharmaz(2000: 511) goes on to situate grounded theory in relation to what was the norm prior to its invention:\n\nGlaser and Strauss’s (1967) work was revolutionary because it challenged (a) arbitrary divisions between theory and research, (b) views of qualitative research as primarily a precursor to more “rigorous” quantitative methods, (c) claims that the quest for rigor made qualitative research illegitimate, (d) beliefs that qualitative methods are impressionistic and unsystematic, (e) separation of data collection and analysis, and (f) assumptions that qualitative research could produce only descriptive case studies rather than theory development (Charmaz 1995).\n\nPrior to Glaser and Strauss (1967), qualitative analysis was taught rather informally — they led the way in providing written guidelines for systematic qualitative data analysis with explicit procedures for data analysis (Charmaz 2000: 512)\nGlaser brought his very positivist assumptions from his work at Columbia, and Strauss’ work in Chicago with Herbert Blumer and Robert Park infused a pragmatic philosophical approach to the study of process, action and meaning that reflects symbolic interactionism.\n\nGlaser\nGlaser’s position comes close to traditional positivism, with assumptions of an objective, external reality and a neutral observer who discovers data. and a reductionist form of inquiry of manageable research problems. According to Charmaz (2000: 511), regarding Glaser’s approach:\n\nTheoretical categories must be developed from analysis of the collected data and must fit them; these categories must explain the data they subsume. This grounded theorists cannot shop their disciplinary stores for preconceived concepts and dress their data in them. Any existing concept must earn its way into the analysis. … The relevance of a grounded theory derives from its offering analytic explanations of actual problems and basic processes in the research setting. A grounded theory is durable because researchers can modify their emerging or established analyses as conditions change or further data are collected.\n\n\n\nCorbin and Strauss\nStrauss and Corbin assume an objective reality, aim toward unbiased data collection, propose a series of technical procedures, and espouses verification. However, they are postpositivism because they propose giving voice to their respondents,3 representing them as accurately as possible, discovering and reckoning with how their respodents’ views on reality differ from their own, and reflecting on the research process as one way of knowing.\nCorbin and Strauss (1990) “gained readers but lost the sense of emergence and open-ended character of Strauss’s earlier volume and much of his empirical work. The improved and more accessible second edition of Basics (Strauss and Corbin 1998) reads as less prescriptive and aims to lead readers to a new way of thinking about their research and about the world.” (Charmaz 2000: 512)\nStrauss apparently became more insistent that grounded theory should be more verificational in nature in personal communications.\nGlaser (1992) responded to Strauss and Corbin (1990), repudiating what he perceived as forcing preconceived questions and frameworks on the data. Glaser considered it better to allow theory to “emerge” from the data, i.e. to let the data speak for themselves.\nCharmaz identifies these two approaches as having a lot in common: hey both advocate for mitigating factors that would hinder objectivity and minimize intrusion of the researcher’s subjectivity, and they are both embedded in positivist attitudes, with a researcher sitting outside the observed reality; Glaser exemplifies these through discovering and coding data, and using systematic comparative methods, whereas Strauss and Corbin maintain a similar distance through their analytical questions, hypotheses and methodological applications. They both engage in “silent authorship” and usually write about their data as distant experts (Charmaz and Mitchell 1996).\n\n\nConstuctivist Grounded Theory\n\nConstructivist grounded celebrates firsthand knowledge of empirical worlds, takes a middle ground between postmodernsm and positivism, and offers accessible methods for taking qualitative research into the 21st century. (510)\n\n\nThe power of grounded theory lies in its tools for understanding empirical worlds. We can reclaim these tools from their positivist underpinnings to form a revised, more open-ended practice of grounded theory that stresses its emergent, constructivist elements. We can use grounded theory methods as flexible, heuristic strategies rather than as formulaic procedures. (510)\n\nThree aspects to Charmaz’s argument (510):4\n\nGrounded theory strategies need not be rigid or prescriptive;\na focus on meaning while using grounded theory furthers, rather than limits, interpretive understanding; and\nwe can adopt grounded theory strategies without embracing the positivist leanings of earlier proponents of grounded theory.\n\nRepudiation of the notion that data speak for themselves, that data do not lie. Recognition that data are constructs of the rsearch process, are framed by the questions we ask informants and the methodological tools of our collection procedures.\nCharmaz (2000: 515) advocates for what seems to be a dialogical approach to coding, between researcher and the data:\n\nWe should interact with our data and pose questions to them while coding. Coding helps us to gain a new perspective on our material and to focus further data collection, and may lead us in unforeseen directions. Unline quantitative research that requires data to fit into preconceived standardized codes, the researcher’s interpretations of data shape his or her emergent codes in grounded theory.\n\nDistinguishes articulates open/initial coding as proceeding line by line to get a general sense of what the data contains. It is meant to keep the researcher close to the data, to remain attuned to the subjects’ views of their realities.\n\nLine-by-line coding sharpens our use of sensitizing concepts — that is, those background ideas that inform the overall research problem. Sensitizing concepts offer eays of seeing, organizing, and understanding experience; they are embedded in our disciplinary emphases and perspectival proclivities. Although sensitizing conceots may deepen perception, they provide starting points for building analysis, not ending points for evading it. We may use sensitizing concepts only as points of departure from which to study the data.\n\nMuch of the rest of the Charmaz (2000) paper is an overview of coding and memoing methods, as well as theoretical sampling. The emphasis is on situating these techniques in the Glaser / Strauss and Corbin debate, and it will be better to refer to Charmaz (2014) for in-depth notes on these techniques.\nCharmaz (2000: 521-522) provides an apt account of a significant critique of grounded theory, and poses her constructivist approach as a potential means of resolving it. Specifically, she refers to the notion that grounded theory (as traditionally conceived by both Glaser and Strauss and Corbin) “fractures” the data, making them easier to digest in an analytical sense, but also making it more difficult to engage with in a holistic manner. This is precisely the point of the original approach, to present qualitative data as data — as conceived and valued by quantitative researchers, i.e. as discrete, corpuscular, disembodied, re-arrangable and distant entities. The text of these two large paragraphs is copied here:\n\nConrad (1990) and Riessman (1990) suggest that “fracturing the data” in grounded theory research might limit understanding because grounded theorists aim for analysis rather than the portrayal of subjects’ experience in its fullness. From a grounded theory perspective, fracturing the data means creating codes and categories as the researcher defines themes within the data. Glaser and Strauss (1967) propose this strategy for several reasons: (a) to help the researcher avoid remaining immersed in anecdotes and stories, and subsequently unconsciously adopting subjects’ perspectives; (b) to prevent the researcher’s becoming immobilized and overwhelmed by voluminous data; and (c) to create a way for the researcher to organize and interpret data. However, criticisms of fracturing the data imply that grounded theory methods lead to separating the experience from the experiencing subject, the meaning from the story, and the viewer from the viewed. In short, the criticisms assume that the grounded theory method (a) limits entry into subjects’ worlds, and thus reduces understanding of their experience; (b) curtails representation of both the social world and subjective experience; (c) relies upon the viewer’s authority as expert observer; and (d) posits a set of objectivist procedures on which the analysis rests.\nResearchers can use grounded theory methods to further their knowledge of subjective experience and to expand its representation while neither remaining external from it nor accepting objectivist assumptions and procedures. A constructivist grounded theory assumes that people create and maintain meaningful worlds through dialectical processes of conferring meaning on their realities and acting within them (Bury 1986; Mishler 1981). Thus social reality does not exist independent of human action. Certainly, my approach contrasts with a number of grounded theory studies, methodological statements, and research texts (see, e.g., Chenitz and Swanson 1986; Glaser 1992; Martin and Turner 1986; Strauss and Corbin 1990; Turner 1981). By adopting a constructivist grounded theory approach, the researcher can move grounded theory methods further into the realm of interpretive social science consistent with a Blumerian (1969) emphasis on meaning, without assuming the existence of a unidimensional external reality. A constructivist grounded theory recognizes the interactive nature of both data collection and analysis, resolves recent criticisms of the method, and reconciles positivist assumptions and postmodernist critiques. Moreover, a constructivist grounded theory fosters the development of qualitative traditions through the study of experience from the standpoint of those who live it.\n\nCharmaz’s (2000: 523) proposal for a re-visioned grounded theory poses research as a materializing process:\n\nA re-visioned grounded theory must take epistemological questions into account. Grounded theory can provide a path for researchers who want to continue to develop qualitative traditions without adopting the positivistic trappings of objectivism and universality. Hence the further development of a constructivist grounded theory can bridge past positivism and a revised future form of interpretive inquiry. A revised grounded theory preserves realism through gritty, empirical inquiry and sheds positivistic proclivities by becoming increasingly interpretive.\n\nCharmaz (2000: 523) addresses realism and truth in constructivist grounded theory, and explicitly relates it to Blumerian situated interactionism:\n\nA constructivist grounded theory distinguishes between the real and the true. The constructivist approach does not seek truth — single, universal, and lasting. Still, it remains realist because it addresses human realities and assumes the existence of real worlds. However, neither human realities nor real worlds are unidimensional. We act within and upon our realities and worlds and thus develop dialectical relations among what we do, think, and feel. The constructivist approach assumes that what we take as real, as objective knowledge and truth, is based upon our perspective (Schwandt 1994). The pragmatist underpinnings in symbolic interactionism emerge here. Thomas and Thomas (1928: 572) proclaim, “If human beings define their situations as real, they are real in their consequences”. Following their theorem, we must try to find what research participants define as real and where their definitions of reality take them. The constructivist approach also fosters our self-consciousness about what we attribute to our subjects and how, when, and why researchers portray these definitions as real. Thus the research products do not constitute the reality of the respondents’ reality. Rather, each is a rendering, one interpretation among multiple interpretations, of a shared or individual reality. That interpretation is objectivist only to the extent that it seeks to construct analyses that show how respondents and the social scientists who study them construct those realities — without viewing those realities as unidimensional, universal, and immutable. Researchers’ attention to detail in the constructivist approach sensitizes them to multiple realities and the multiple viewpoints within them; it does not represent a quest to capture a single reality.\nThus we can recast the obdurate character of social life that Blumer (1969) talks about. In doing so, we change our conception of it from a real world to be discovered, tracked, and categorized to a world made real in the minds and through the words and actions of its members. Thus the grounded theorist constructs an image of a reality, not the reality — that is, objective, true, and external.\n\nOn the other hand, Charmaz (2000: 524) frames objectivist grounded theory as believing in some kind of truth:\n\nObjectivist grounded theory accepts the positivistic assumption of an external world that can be described, analyzed, explained, and predicted: truth, but with a small t. That is, objectivist grounded theory is modifiable as conditions change. It assumes that different observers will discover this world and describe it in similar ways That’s correct — to the extent that subjects have comparable experiences (e.g., people with different chronic illnesses may experience uncertainty, intrusive regimens, medical dominance) and viewers bring similar que-tions, perspectives, methods, and, subsequently, concepts to analyze those experiences. Objectivist grounded theorists often share assumptions with their research participants — particularly the professional participants. Perhaps more likely, they assume that respondents share their meanings. For example, Strauss and Corbin’s (1990) discussion of independence and dependence assumes that these terms hold the same meanings for patients as for researchers.\n\nCharmaz (2000: 525) further embeds construvist grounded theory as a way to fulfill Blumer’s symbolic interactionism:\n\nWhat helps researchers develop a constructivist grounded theory? How might they shape the data collection and analysis phases? Gaining depth and understanding in their work means that they can fulfill Blumer’s (1969) call for “intimate familiarity” with respondents and their worlds (see also Lofland and Lofland 1984, 1995). In short, constructing constructivism means seeking meanings — both respondents’ meanings and researchers’ meanings.\n\nCharmaz (2000: 524) on the concretization of procedures from what were orginally meant to be guidelines:\n\nGuidelines such as those offered by Strauss and Corbin (1990) structure objectivist grounded theorists’ work. These guidelines are didactic and prescriptive rather than emergent and interactive. Sanders (1995: 92) refers to grounded theory procedures as “more rigorous than thou instructions about how information should be pressed into a mold”. Strauss and Corbin categorize steps in the process with scientific terms such as axial coding and conditional matrix (Strauss 1987; Strauss and Corbin 1990, 1994). As grounded theory methods become more articulated, categorized, and elaborated, they seem to take on a life of their own. Guidelines turn into procedures and are reified into immutable rules, unlike Glaser and Strauss’s (1967) original flexible strategies. By taking grounded theory methods as prescriptive scientific rules, proponents further the positivist cast to obiectivist grounded theory.\n\n\nOn the modes of reasoning behind grounded theory\nKelle (2005) is an overview of the Glaser / Strauss and Corbin split. References to Kelle (2005) have no page numbers since it is published in an online-only journal and does not specify paragraph numbers.\nHighlights a primary impetus behind Glaser and Strauss (1967), which used political analogies to distinguish between “theoretical capitalists” and “proletariat testers”, and unify the field of sociology by de-centering emphasis on theories developed by “great men”.\nA common thread in this paper is sensitivity to the practical challenges of actually doing grounded theory according to Glaser’s approach:\n\nThe infeasibility of an inductivist research strategy which demands an empty head (instead of an “open mind”) cannot only be shown by epistemological arguments, it can also be seen in research practice. Especially novices in qualitative research with the strong desire to adhere to what they see as a basic principle and hallmark of Grounded Theory — the “emergence” of categories from the data — often experience a certain difficulty: in open coding the search for adequate coding categories can become extremely tedious and a subject of sometimes numerous and endless team sessions, especially if one hesitates to explicitly introduce theoretical knowledge. The declared purpose to let codes emerge from the data then leads to an enduring proliferation of the number of coding categories which makes the whole process insurmountable.\n\nKelle (2005) basically takes down the original Glaser and Strauss (1967) and subsequent reflection on theoretecal sensitivity (Glaser 1978). He highlights fundamental contraditions and oversights with regards to the role of theory in grounded theory, specifically with regards to the notion that such research can be accomplished with inductive purity:\n\nConsequently, in the most early version of Grounded Theory the advice to employ theoretical sensitivity to identify theoretical relevant phenomena coexists with the idea that theoretical concepts “emerge” from the data if researchers approach the empirical field with no preconceived theories or hypotheses. Both ideas which have conflicting implications are not integrated with each other in the Discovery book. Furthermore, the concept of theoretical sensitivity is not converted into clear cut methodological rules: it remains unclear how a theoretically sensitive researcher can use previous theoretical knowledge to avoid drowning in the data. If one takes into account the frequent warnings not to force theoretical concepts on the data one gets the impression that a grounded theorist is advised to introduce suitable theoretical concepts ad hoc drawing on implicit theoretical knowledge but should abstain from approaching the empirical data with ex ante formulated hypotheses.\n\nKelle (2005) recognizes that Glaser identified a series of “theoretical families” to help assist with the practical experience of coding. I find it somewhat interesting that many of the terms in these first families are very reminiscient of so-called “natural language”, as used in the wave of cybernets that was contemporary with Glaser (1978) and which largely dealt with “expert systems”.\n\nIn the book “Theoretical Sensitivity” (1978) GLASER presents an extended list of terms which can be used for the purpose of theoretical coding loosely structured in the form of so called theoretical “coding families”. Thereby various theoretical concepts stemming from different (sociological, philosophical or everyday) contexts are lumped together, as for example:\n\nterms, which relate to the degree of an attribute or property (“degree family”), like “limit”, “range”, “extent”, “amount” etc.,\nterms, which refer to the relation between a whole and its elements (“dimension family”), like “element”, “part”, “facet”, “slice”, “sector”, “aspect”, “segment” etc.,\nterms, which refer to cultural phenomena (“cultural family”) like “social norms”, “social values”, “social beliefs” etc.\n\n\nThis is substantiated by other observations by Kelle (2005) that ad hoc coding actually follows implicit theoretical knowledge:\n\nOne of the most crucial differences between GLASER’s and STRAUSS’ approaches of Grounded Theory lies in the fact that STRAUSS and CORBIN propose the utilization of a specified theoretical framework based on a certain understanding of human action, whereas GLASER emphasises that coding as a process of combining “the analyst’s scholarly knowledge and his research knowledge of the substantive field” (1978, p.70) has to be realised ad hoc, which means that it has often to be conducted on the basis of a more or less implicit theoretical background knowledge.\n\nand that the Glaserian approach is better suited for more experienced, rather than novice sociologists, who will have internalized the theory that they then apply in their coding.\nKelle then goes on to address how grounded theory can or can not be applied in alignment with inductivist or hypothetic-deductivist reasoning, and raises abductive reasoning an an alternative means of arriving at legitimate and verifiable conclusions. There is too much detail in the paper to copy here.\nBut here is another nice conclusive gem from the end:\n\nWhereas STRAUSS and CORBIN pay a lot of attention to the question how grounded categories and propositions can be further validated, GLASER’s concept shows at least a gleam of epistemological fundamentalism (or “certism”, LAKATOS 1978) especially in his defence of the inductivism of early Grounded Theory. “Grounded theory looks for what is, not what might be, and therefore needs no test” (GLASER 1992, p.67). Such sentences carry the outmoded idea that empirical research can lead to final certainties and truths and that by using an inductive method the researcher may gain the ability to conceive “facts as they are” making any attempt of further corroboration futile.\n\n\n\nRebuttals by Glaser\nGlaser (2002) constitutes a rebuttal to Charmaz (2000). As Bryant (2003) points out in his response to Glaser (2002), it is very angry, polemical and irrational. I don’t want to go too in depth with the fundamental problems with Glaser’s response (see Bryant’s paper for the details), but the gist is that Glaser never really got the message about data being inherently constructed by researchers decisions, actions and circumstances. Glaser seems to continue believing in the inherent neutrality of data as a matter of faith.\nThis being said, Glaser (2002) did highlight the large emphasis on descriptive rather than explanatory potential in Charmaz’s approach. This aligns with my own apprehensions when I try to address the relevance of my work. I tend to use the term “articulate” as a way to frame my work as descriptive, but in a way that lends value, and this very fuzzy distinction between the power of identying the shapes and relationships among things and explaining their causes and effects in a generalizable way (i.e., theories, or explanations), still somehow troubles me. I wonder if Glaser is drawing a false distinction here, and through that, a false prioritization of explanation over description as a desired outcome. This would put my mind at ease, as would dismissing Glaser’s dismissal of people who simply don’t know how to do the “real” grounded theory (which, in his mind, include all feminist and critical researchers).\n\n\nOn the utility of grounded theory\nI completely agree with this statement from Clarke (2003: 555):\n\nTo address the needs and desires for empirical understandings of the complex and heterogeneous worlds emerging through new world orderings, new methods are requisite (Haraway 1999). I believe some such methods should be epistemologically/ ontologically based in the pragmatist soil that has historically nurtured symbolic interactionism and grounded theory. Through Mead, an interactionist grounded theory has always had the capacity to be distinctly perspectival in ways fully com patible with what are now understood as situated knowledges. This fundamental and always already postmodern edge of a grounded theory founded in symbolic interactionism makes it worth renovating.\n\nThis is super interesting, and really contextualizes how Strauss imagined grounded theory to be useful for him:\n\nSome years ago, Katovich and Reese (1993:400–405) interestingly argued that Strauss’s negotiated order and related work recuperatively pulled the social around the postmodern turn through its methodological [grounded theoretical] recognition of the partial, tenuous, shifting, and unstable nature of the empirical world and its constructedness. I strongly agree and would argue that Strauss also furthered this “postmodernization of the social” through his conceptualizations of social worlds and arenas as modes of understanding the deeply situated yet always also fluid orga nizational elements of negotiations. He foreshadowed what later came to be known as postmodern assumptions: the instability of situations; characteristic changing, porous boundaries of both social worlds and arenas; social worlds seen as mutually constitutive/coproduced through negotiations taking place in arenas; negotiations as central social processes hailing that “things can always be otherwise”; and so on. Significantly, negotiations constitute discourses that also signal micropolitics of power as well as “the usual” meso/macrostructural elements—power in its more fluid forms (e.g., Foucault 1979, 1980). Through integrating the social worlds/arenas/ negotiations framework with grounded theory as a new conceptual infrastructure, I hope to sustain and extend the methodological contribution of grounded theory to understanding and elaborating what has been meant by “the social” in social life — before, during, and after the postmodern turn.\n\nIt also echoes Charmaz’s vision of grounded theory as a powerful too, and Bryant’s (2003) call to “look at what Glaser and Strauss actually did, rather than what they claimed — and continued to claim — they were doing” to uncover “the basis for a powerful research approach”. Bryant (2003) further cites Baszanger and Dodier (1997), who characterize grounded theory as a method “consisting of accumulating a series of individual cases, of analyzing them as a combination between different logics of action that coexist not only in the field under consideration, but even within these individuals or during their encounters”. Bryant (2003) summarizes this by stating that “[t]he aim of such methods is generalization rather than totalization, with the objective of producing”a combinative inventory of possible situations”." + "objectID": "notes/maelstrom-readings.html#doiron2017", + "href": "notes/maelstrom-readings.html#doiron2017", + "title": "Maelstrom reading notes", + "section": "Doiron et al. (2017)", + "text": "Doiron et al. (2017)\n\nAn overview of the key software that facilitates data harmonization practices under Maelstrom, also briefly touched upon in Doiron et al. (2013).\nPage 1373 refers to graphical and programmatic interfaces and assumes certain roles and tasks associated with each.\nBriefly describes its use by the Canadian Longitudinal Study on Aging (CLSA), the Canadian Partnership for Tomorrow Project (CPTP) and InterConnect, primarily by describing the range and quantity of data that these systems manage in each case.\n\n\nOpal provides a centralized web-based data management system allowing study coordinators and data managers to securely import/export a variety of data types (e.g. text, nu merical, geolocation, images, videos) and formats (e.g. SPSS, CSV) using a point-and-click interface. Opal then converts, stores and displays these data under a standar dized model.\n\n\nMica is used to create websites and metadata portals for individual epidemiological studies or multi-study consor tia, with a specific focus on supporting observational co hort studies. The Mica application helps data custodians and study or network coordinators to efficiently organize and disseminate information about their studies and net works without significant technical effort." }, { - "objectID": "notes/methodology-notes.html#theoretical-sampling", - "href": "notes/methodology-notes.html#theoretical-sampling", - "title": "Methodology notes", - "section": "Theoretical sampling", - "text": "Theoretical sampling\nSee Charmaz (2000): 519-520.\nFrom Clarke (2003: 557):\n\nUnique to this approach has been, first, its requiring that analysis begin as soon as there are data. Coding begins immediately, and theorizing based on that coding does as well, however provisionally (Glaser 1978). Second, “sampling” is driven not necessarily (or not only) by attempts to be “representative” of some social body or population (or its heterogeneities) but especially and explicitly by theoretical con cerns that have emerged in the provisional analysis. Such “theoretical sampling” focuses on finding new data sources (persons or things) that can best explicitly ad dress specific theoretically interesting facets of the emergent analysis. Theoretical sampling has been integral to grounded theory from the outset, remains a fundamen tal strength of this analytic approach, and is crucial for the new situational analyses." + "objectID": "notes/maelstrom-readings.html#fortier2010", + "href": "notes/maelstrom-readings.html#fortier2010", + "title": "Maelstrom reading notes", + "section": "Fortier et al. (2010)", + "text": "Fortier et al. (2010)\n\nA very grandiose paper presenting the grand vision for DataSHaPER, which would eventually become Maelstrom.\n\nLots of co-authors!\n\nInvokes the pan-European EPIC project (European Prospective Investigation into Cancer and Nutrition), which faced numerous data synthesis challenges despite its proactive effort to coordinate work across numerous research centres.\n\n\nTwo complementary approaches may be adopted to support effective data synthesis. The first one principally targets ‘what’ is to be synthesized, whereas the other one focuses on ‘how’ to collect the required information. Thus: (i) core sets of information may be identified to serve as the foundation for a flexible approach to harmonization; or (ii) standard collection devices (questionnaires and stand ard operating procedures) may be suggested as a required basis for collection of information.\n\n\nDataSHaPER is an acronym for DataSchema and Harmonization Platform for Epidemiological Research.\n\n\nIn an ideal world, information would be ‘prospectively harmonized’: emerging studies would make use, where possible, of harmonized questionnaires and standard operating procedures. This enhances the potential for future pooling but entails significant challenges —- ahead of time -— in developing and agree ing to common assessment protocols. However, at the same time, it is important to increase the utility of existing studies by ‘retrospectively harmonizing’ data that have already been collected, to optimize the subset of information that may legitimately be pooled. Here, the quantity and quality of infor mation that can be pooled is limited by the heterogeneity intrinsic to the pre-existing differences in study design and conduct.\n\nCompares prospective and retrospective harmonizatiom, with the former being presented as ideal, and the latter being a pragmatic reconciliation in acknowledgement that the former is essentially impossible to achieve.\n\nDataSHaPER is strikingly similar to OCHRE:\n\nXML-based data structures\nGenesis of a generic and ultimately optional base-level schema that illustrates the kind of data that the data structure may hold in ways that are immediately recognizable to all practitioners (at OCHRE it was associations between contexts and finds)\nSeparate harmonization platform where users can edit and manipulate records and associations between them\n\n\n\nThe question ‘What would constitute the ultimate proof of success or failure of the DataSHaPER approach’ needs to be addressed. Such proof will necessarily accumulate over time, and will involve two fundamental elements: (i) ratification of the basic DataSHaPER approach; and (ii) confirmation of the quality of each individual DataSHaPER as they are developed and/or extended. An important indication of the former would be provided by the widespread use of our tools. However, the ultimate proof of principle will necessarily be based on the generation of replicable scientific findings by researchers using the approach. But, for such evidence to accumulate it will be essential to assure the quality of each individual DataSHaPER. Even if the fundamental approach is sound, its success will depend critically on how individual DataSHaPERs are constructed and used. It seems likely that if consistency and quality are to be assured in the global development of the approach, it will be necessary for new DataSHaPERs to be formally endorsed by a central advisory team." }, { - "objectID": "notes/methodology-notes.html#interviews", - "href": "notes/methodology-notes.html#interviews", - "title": "Methodology notes", - "section": "Interviews", - "text": "Interviews\nSee (Yin 2014: 110-113) See Becker (1998)\nFrom Charmaz (2000: 525):\n\nA constructivist approach necessitates a relationship with respondents in which they can cast their stories in their terms. It means listening to their stories with openness to feeling and experience. … Furthermore, one-shot interviewing lends itself to a partial, sanitized view of experience, cleaned up for public discourse. The very structure of an interview may preclude private thoughts and feelings from emerging. Such a structure reinforces whatever proclivities a respondent has to tell only the public version of the story. Researchers’ sustained involvement with research participants lessens these problems.\n\nFontana and Frey (2000) spend some time writing about the emergence of an “interview society”, whereby interviews are commonly used to seek various forms of biographical information. They cite Gubrium and Holstein (1998), who noted that “the interview has become a means of contemporary storytelling, where persons divulge life accounts in response to interview inquiries”. They then go over a brief history of interviewing in the context of sociological research, which largely tracks the values underlying positivist and postmodernist transitions as you might expect.\n\nStructured interviewing\nFrom Fontana and Frey (2000: 649-651):\nInterviewers ask respondents a series of preestablished questions with a limited set of response categories. The interview records responses according to a preestablished coding scheme.\nInstructions to interviewers often follow these guidelines:\n\nNever get involved in long explanations of the study; use the standard explanation provided by the supervisor.\nNever deviate from the study introduction, sequence of questions, or question wording.\nNever let another person interrupt the interview; do not let another person answer for the respondent or offer his or her opinions on the question.\nNever suggest an answer or agree or disagree with an answer. Do not give the respondent any idea of your personal views on the topic of the question or the survey.\nNever interpret the meaning of a question; just repeat the question and give instructions or clarifications that are provided in training or by the supervisors.\nNever improvise, such as by assing answer categories or making wording changes.\n\nThe interviewer must establish a “balanced rapport”, being casual and friendly while also directive and impersonal. Interviewers must also perfect a style of “interested listening” that rewards respondents’ participation but does not evaluate their responses.\nFrom Fontana and Frey (2000: 651):\n\nThis kind of interview often elicits rational responses, but it overlooks or inadequately assesses the emotional dimension.\n\n\n\nGroup interviews\nFrom Fontana and Frey (2000: 651-652):\nCan be used to test a methodological technique, try out a definition of a research problem or to identify key informants. Pre-testing a questionnaire or survey design.\nCan be used to aid respondents’ recall of specific events or to stimulate embellished descriptions of events, or experiences shared by members of a group.\nIn formal group interviews, participants share views through the coordinator.\nLess formal group interviews are meant to establish the widest range of meaning and interpretation on a topic, and the objective is “to tap intersubjective meaning with depth and diversity”.\n\n\nUnstructured interviewing\nFrom Fontana and Frey (2000: 652-657):\nThe essence of an unstructured interview is establishing a human-to-human relation with the respondent and a desire to understand rather than to explain.\nFontana and Frey (2000) then goes on with some practical guidance on how to engage in unstructured interviews, largely concerned with how to access a community and relate with respondents.\n\n\nPostmodern takes on interviewing\nFontana and Frey (2000) address some “new” takes on interviewing emerging from the postmodern turn. I kinda think there is some potential behind approaches that emphasize interviews as negotiated accomplishment, or product of communal sensemaking between interviewer and respondent. I think it could be really helpful in the context of my research, which is very concerned with drawing out tensions that respondents have in mind but are not really able to articulate in a systematic way.\nHowever, Fontana and Frey (2000) also draws attention to crticism of highly engaged interviewing approaches, which seem to equate closeness with the respondent as getting closer to their “true self”, and which may not actually be fixed (especially in the context of the artificial environment of the interview setting). Critiques of such “closeness” use the term “romantic” or “crusading” as epithets. Moreover, there is additional reference to the culturally-embedded assumption that interviews are necessarily valuable sources of information, as if speaking ones mind can adequately convey one’s thoughts and experiences — this is criticized as a particularly western approach to information extraction surrounding internalized and externalized thoughts and behaviour, as instilled through participation in the “interview society” addressed earlier in the text.\n\n\nTranscribing\nThis section describes how I transcibe interviews and accounts for the decisions to encode certain things and not others. It goes on to explains the procedures for transcribing spoken dialog into textual formats, including the notation applied to encode idiosyncratic elements of conversational speech.\n\nTranscript notation\nDerived from the transcription protocol applied for the E-CURATORS project.\n\n\nCleaning audio\nTo clean the audio:\n\nSelect a clip that is representative of a single source of background noise, and then filter that wavelength throughout the entire audio file.\nAfter selecting the clip, go to Effect >> Noise Reduction and select Get Noise Profile, then press OK.\nClose the noise reduction menu, select the entire range of audio using the keyboard shortcut Command + A.\nThen go back to the noise reduction window (Effect >> Noise Reduction) to apply the filter based on the noise profile identified for the noisy clip.\nExport the modified audio file to the working directory (File >> Export >> Export as .WAV).\nUse ffmpeg to replace the dirty audio track with the clean one:\n\n ffmpeg -i dirty.mp4 -i clean.wav -c:v copy -map 0:v:0 -map 1:a:0 clean.mp4" + "objectID": "notes/maelstrom-readings.html#fortier2011", + "href": "notes/maelstrom-readings.html#fortier2011", + "title": "Maelstrom reading notes", + "section": "Fortier et al. (2011)", + "text": "Fortier et al. (2011)\nThis paper responds to Hamilton et al. (2011), which presents an effort to devise a standardized nomenclature. The response is basically to advocate for a more flexible approach, rather than a stringent one promoted by Hamilton et al. (2011). It draws extensively from concepts published in the foundational paper by Fortier et al. (2010).\n\nTwo complementary approaches to harmonization may be adopted to support effective data synthesis or comparison across studies. The first approach makes use of identical data collection tools and procedures as a basis for harmoni zation and synthesis. Here we refer to this as the ‘‘stringent’’ approach to harmonization. The second approach is con sidered ‘‘flexible’’ harmonization. Critically, the second ap proach does not demand the use of identical data collection tools and procedures for harmonization and synthesis. Rather, it has to be based on sound methodology to ensure inferential equivalence of the information to be harmonized. Here, standardization is considered equivalent to stringent harmonization. It should, however, be noted that the term standard is occasionally employed to refer to common con cepts or comparable classification schemes but does not necessarily involve the use of identical data collection tools and procedures (12, 13).\n\nThis directly parallels the distinction made in Fortier et al. (2010) between “ideal” prospective and more pragmatic retrospective approaches to data harmonization.\n\nSynthesis of data using a flexible harmonization approach may be either prospective or retrospective. To achieve flexible prospective harmonization, investigators from several studies will agree on a core set of variables (or measures), compatible sets of data collection tools, and standard operating procedures but will allow a certain level of flexibilit in the specific tools and procedures used in each study (16, 17). Retrospective harmonization targets synthesis of information already collected by existing legacy studies (15, 18, 19). As an illustrative example, using retrospective harmonization, researchers will define a core set of variables (e.g., body mass index, global level of physical activity) and, making use of formal pairing rules, assess the potential for each participating study to create each variable (15). The ability to retrospectively harmonize data from existing studies facilitates the rapid generation of new scientifi knowledge.\n\nI wonder why there is no example provided for prospective data harmonization. Is it because it is ideal and not realistic? I’d argue that it is simply what occurs within individual projects." }, { - "objectID": "notes/methodology-notes.html#field-notes", - "href": "notes/methodology-notes.html#field-notes", - "title": "Methodology notes", - "section": "Field notes", - "text": "Field notes\nSee (Yin 2014: 124-125)" + "objectID": "notes/maelstrom-readings.html#fortier2017", + "href": "notes/maelstrom-readings.html#fortier2017", + "title": "Maelstrom reading notes", + "section": "Fortier et al. (2017)", + "text": "Fortier et al. (2017)\nExplicit statement regarding the rationale and presumed benefits of harmonization right in the first paragraph:\n\nThe rationales underpinning such an approach include ensuring: sufficient statistical power; more refined subgroup analysis; increased exposure hetero geneity; enhanced generalizability and a capacity to under take comparison, cross validation or replication across datasets. Integrative agendas also help maximizing the use of available data resources and increase cost-efficiency of research programmes.\n\nSummarized in bullet points:\n\nensuring sufficient statistical power\nmore refined subgroup analysis\nincreased exposure heterogeneity\nenhanced generalizability\na capacity to undertake comparison, cross validation or replication across datasets.\nmaximizing the use of available data resources\nincrease cost-efficiency of research programmes\n\nClearly defines harmonization and its benefits:\n\nEssentially, data harmonization achieves or improves comparability (inferential equivalence) of similar measures collected by separate studies.\n\nAdds an additional argument for retrospective harmonization on top of prior discussion of retrospective/prospective approaches (cf. Fortier et al. (2010); Fortier et al. (2011)):\n\nRepeating identical protocols is not necessarily viewed as providing evidence as strong as that obtained by exploring the same topic but using different designs and measures.\n\nAlso relates retrospective harmonization from systematic meta reviews. In fact, the paper basically responds to calls for more structured guidelines for data harmonization, similar to those that had been produced to support structured metareviews in the years prior to this publication. The authors identify several papers that have done similar guidelines or reports on harmonization practices, which they claim are too broad. Those papers include:\n\nRolland et al. (2015)\n\nhttps://doi.org/10.1093/aje/kwv133\n\nSchaap et al. (2011)\n\nhttps://doi.org/10.1186/1471-2474-12-272\n\nBennett et al. (2011)\n\nhttps://doi.org/10.1002/gepi.20564\n\nHohmann et al. (2012)\n\nThis paper reports the findings of a systematic inquiry made to data harmonization initiatives, whose data comprise responses to a questionnaire. The findings indicate that procedures were more attentively follows during earlier stages, such as when matching and aligning available data with the project’s designated scope. However, procedures were less sound with regards to documenting procedures, validating the results of data processing, and dissemination strategy. There is a notable division between work that occurs before and after people actually begin handling the data, which indicates a tension between aspirational plans and tangible experiences.\n\nRespondents were asked to delineate the specific procedures or steps undertaken to generate the harmonized data requested. Sound procedures were generally described; however, the terminologies, sequence and technical and methodological approaches to these procedures varied considerably. Most of the procedures mentioned were related to defining the research questions, identifying and selecting the participating studies (generally not through a systematic approach), identifying the targeted variables to be generated and processing data into the harmonized variables. These procedures were reported by at least 75% of the respondents. On the other hand, few reported steps related to validation of the harmonized data (N=4; 11.8%), documentation of the harmonization process (N=5; 14.7%) and dissemination of the harmonized data outputs (N=2; 5.9%).\n\nThe paper summarizes some specific “potential pitfalls” reported by respondents to their survey:\n\nensuring timely access to data;\nhandling dissimilar restrictions and procedures related to individual participant data access;\nmanaging diversity across the rules for authorship and recognition of input from study-specific investigators;\nmobilizing sufficient time and resources to conduct the harmonization project;\ngathering information and guidance on harmonization approaches, resources and techniques;\nobtaining comprehensive and coherent information on study-specific designs, standard operating procedures, data collection devices, data format and data content;\nunderstanding content and quality of study-specific data;\ndefining the realistic, but scientifically acceptable, level of heterogeneity (or content equivalence) to be obtained;\ngenerating effective study-specific and harmonized datasets, infrastructures and computing capacities;\nprocessing data under a harmonized format taking into account diversity of: study designs and content, study population, synchronicity of measures (events measured at different point in time or at different intervals when repeated) etc;\nensuring proper documentation of the process and decisions undertaken throughout harmonization to ensure transparency and reproducibility of the harmonized datasets;\nmaintaining long-term capacities supporting dissemination of the harmonized datasets to users.\n\nIt’s not made clear how these responses were distributed among respondents.\nThe authors then identify several absolute essential requirements needed to achieve success:\n\nCollaborative framework: a collaborative environment needs to be implemented to ensure the success of any harmonization project. Investigators involved should be open to sharing information and knowledge, and investing time and resources to ensure the successful implementation of a data-sharing infrastructure and achievement of the harmonization process.\nExpert input: adequate input and oversight by experts should be ensured. Expertise is often necessary in: the scientific domain of interest (to ensure harmonized variables permit addressing the scientific question with minimal bias); data harmonization methods (to support achievement of the harmonization procedures); and ethics and law (to address data access and integration issues).\nValid data input: study-specific data should only be harmonized and integrated if the original data items collected by each study are of acceptable quality.\nValid data output: transparency and rigour should be maintained throughout the harmonization process to ensure validity and reproducibility of the harmonization results and to guarantee quality of data output. The common variables generated necessarily need to be of acceptable quality.\nRigorous documentation: publication of results generated making use of harmonized data must provide the information required to estimate the quality of the process and presence of potential bias. This includes a description of the: criteria used to select studies; process achieved to select and define variables to be harmonized; procedures used to process data; and characteristics of the study-specific and harmonized dataset(s) (e.g. attribute of the populations).\nRespect for stakeholders: all study-specific as well as network-specific ethical and legal components need to be respected. This includes respect of the rights, intellectual property interests and integrity of study participants, investigators and stakeholders.\n\nThe authors describe how they arrived at guidelines following the results of this study:\n\nA consensus approach was used to assemble information about pitfalls faced during the harmonization process, establish guiding principles and develop the guidelines. The iterative process (informed by workshops and case studies) permitted to refine and formalize the guide lines. The only substantive structural change to the initial version proposed was the addition of specific steps relating to the validation, and dissemination and archiving of harmonized outputs. These steps were felt essential to em phasize the critical nature of these particular issues.\n\nThe paper outlines a checklist of stages that data harmonization initiatives need to go through to produce ideal outcomes. For each task, they describe a scenario in which the task can be said to be complete, whhich resembles an ideal outcome. This is described in the paper, summarized in a table, and more comprehensively documented in the supplementary materials.\nAlso worth noting, this paper includes a list of harmonization initiatives that I may consult when selecting cases. I’m not quite sure how useful it will be since the findings don’t really break down the distribution of responses in any detail, but maybe the authors have done this analysis and not published it." }, { - "objectID": "notes/methodology-notes.html#focus-groups", - "href": "notes/methodology-notes.html#focus-groups", - "title": "Methodology notes", - "section": "Focus groups", - "text": "Focus groups\nFrom Kitzinger (1994); Wilkinson (1998); Parker and Tritter (2006); Morgan and Hoffman (2018).\n\nCrucially, focus groups involve the interaction of group participants with each other as well as with the moderator, and it is the collection of this kind of interactive data which distinguishes the focus group from the one-to-one interview (c.f. Morgan 1988), as well as from procedures which use multiple participants but do not permit interactive discussion (c.f. Stewart and Shamdasani 1990). The ‘hallmark’ of focus groups, then, is the ‘explicit use of group interaction to produce data and insights that would be less accessible without the interaction found in a group’ (Morgan 1997: 2).\n\nWilkinson (1998) identifies three key features of focus group methods:\n\nproviding access to participants’ own language, concepts and concerns;\n\n\n“The relatively free flow of discussion and debate between members of a focus group offers an excellent opportunity for hearing ‘the language and vernacular used by respondents’”\nUseful for gaining insight into participants’ conceptual worlds on their own terms\n“Sometimes, the participants even offer the researcher a ‘translation’ of unfamiliar terms or concepts”\n“Focus group interactions reveal not only shared ways of talking, but also shared experiences, and shared ways of making sense of these experiences. The researcher is offered an insight into the commonly held assumptions, concepts and meanings that constitute and inform participants’ talk about their experiences.”\n\n\nencouraging the production of more fully articulated accounts;\n\n\n“In focus groups people typically disclose personal details, reveal discrediting information, express strong views and opinions. They elaborate their views in response to encouragement, or defend them in the face of challenge from other group members: focus groups which ‘take off’ may even, like those run by Robin Jarrett (1993: 194) have ‘the feel of rap sessions with friends’.”\n“Even when bcus group participants are not acquainted in advance, the interactive nature of the group means that participants ask questions of, disagree with, and challenge each other, thus serving ‘to elicit the elaboration of responses’ (Merton 1987: 555).”\n“Other ethical issues in focus group research stem from group dynamics, insofar as participants can collaborate or collude effectively to intimidate and/or silence a particular member, or to create a silence around a particular topic or issue, for example. In such cases, it falls to the group moderator to decide whether/how to intervene, and it can be difficult to balance such conflicting goals as ensuring the articulation of accounts, supporting individuals, and challenging offensive statements.”\n\n\nand offering an opportunity to observe the process of collective sense-making.\n\n\nFocus groups also offer an opportunity for researchers to see exactly how views are constructed, expressed, defended and (sometimes) modified during the course of conversations with others, 1.e. to observe the process of collective sense-making in action. This is likely to be of particular interest to researchers working within a social constructionist framework, who do not view beliefs, ideas, opinions and understandings as generated by individuals in splendid isolation, but rather as built in interaction with others, in specific social contexts: as Radley and Billig (1996: 223) say, ‘thinking is a socially shared activity’. In a focus group, people are confronted with the need to make collective sense of their individual experiences and beliefs (Morgan and Spanish 1984: 259).\n\nWilkinson (1998) addresses opportunities and challenges in analyzing focus group data using content analysis and ethnographic techniques.\nWith regards to content analysis:\n\nthe main advantages of content analysis are to allow for a relatively systematic treatment of the data and to enable its presentation in summary form. … the researcher has first to decide on the unit of analysis: this could be the whole group, the group dynamics, the individual participants, or the participants’ utterances\n\n\nMorgan (1997) proposes three distinct ways of coding focus group data: noting whether each group discussion contains a given code; noting whether each participant mentions a given code; and noting all mentions of a given code (i.e. across groups or participants). Once the data have been coded in one (or more) of these ways, the question of whether to quantify them is a further issue. Morgan (1993) argues the value of simple ‘descriptive counts’ of codes (stopping short of using inferential statistical tests, whose assumptions are unlikely to be met in focus groups)\n\nAnd with regards to ethnographic analysis:\n\nIts main advantage is to permit a detailed interpretative account of the everyday social processes of communication, talk and action occurring within the focus group. The key issue in ethnographic analysis is how to select the material to present (whether this is framed up as ‘themes’, ‘discourses’, or simply as illustrative quotations), without violating the ‘spirit’ of the group, and without losing sight of the specific context within which the material was generated. … A particular challenge is how to preserve the interactive nature of focus group data: a surprising limitation of published focus group research is the rarity with which group interactions are analysed and reported ( c.f. Carey and Smith 1994, Kitzinger 1994a). Extracts from focus group data are most commonly presented as if they were one-to-one interview data, often with no indication that more than one person is present; still more rarely does interaction per se constitute the analytic focus.\n\n\nKitzinger (1994) elaborates on the interactive nature of focus groups, which she identifies as the methods’ core feature.\n\nFocus groups are group discussions organised to explore a specific set of issues sucb as people’s views and experiences … The group is ‘focused’ in the sense that it involves some kind of collective activity - such as viewing a film, examining a single health education message or simply debating a particular set of questions.\n\n\nEven when group work is explicitly included as part of the research it is often simply employed as a convenient way to illustrate a theory generated by other methods or as a cost-effective technique for interviewing several people at once. Reading some such reports it is hard to believe that there was ever more than one person in the room at the same time. This criticism even applies to many studies which explicitly identify their methodology as ‘focus group discussion’ — in spite of the fact that the distinguishing feature of focus groups is supposed to be the use of interaction as part of the research data.\n\n\nIt would be naive, however, to assume that group data is by definition ‘natural’ in the sense that it would have occurred without the group having been convened for this purpose. It is important to note that although, at times, the focus groups may approximate to participant observation the focus groups are artifidally set up situations. Rather than assuming the group session unproblematically and inevitably reflects ‘everyday interactions’ (although sometimes it will) the group sbould be used to encourage people to engage with one another, verbally formulate their ideas and draw out the cognitive structures which previously have been unarticulated.\n\nOn the active role of the facilitator:\n\nSessions were conducted in a relaxed fashion with minimal intervention from the facilitator - at least at first. This allowed the facilitator to ‘find her feet’ and permitted the research participants to set the priorities. However, the researcber was never passive. Trying to maximise interaction between participants could lead to a more interventionist style: urging debate to continue beyond the stage it might otberwise have ended, challenging people’s taken for granted reality and encouraging them to discuss the inconsistencies both between participants and within their own thinking.\n\nOn the role of the presentation or activity:\n\nSuch exercises not only provided invaluable data from each group but allow for some cross-compadsons between groups. Each discussion session has its own dynamic and direction — when it comes to analysis it is extremely useful to have a common external reference point such as that provided by the card game or the use of vignettes (Kban and Manderson 1992).\n\nOn complementary interactions:\n\nThe excbange between the research participants not only allows the researcher to understand which advertisement they are talking about but to gather data on their shared perception of that image.\n\n\nBrainstorming and loose word association was a frequent feature of the research sessions.\n\n\npeople’s knowledge and attitudes are not entirely encapsulated in reasoned responses to direct questions. Everyday forms of communication such as anecdotes, jokes or loose word association may tell us as much, if not more, about what people ‘know’. In this sense focus groups ‘reach the parts that other methods cannot reach’ - revealing dimensions of understanding that often remain untapped by the more conventional one-to-one interview or questionnaire.\n\n\nIn addition to the advantages discussed above fo exampie, discussing whether they had the ‘right’ to know if another child in the piay group bad had virus asserted that ‘you think of your own first’. It was this phrase, and these sort of sentiments, which seemed to capture their consent and resulted in nods of agreement round the group and assertions that ‘that’s right’ and ‘of course’. Indeed, it was often the strength of the collective reaction that highlighted the specific context within which the research participants experienced AIDS information.\n\nOn argumentative interactions:\n\nthe group process however, is not only about consensus and the articulation of group norms and experiences. Differences between individuals within the group are equally important and, in any case, rarely disappear from view. Regardless of how they are selected, the research participants in any one group are never entirely homogenous. Participants do not just agree with each other — they also misunderstand one another, question one another, try to persuade each other of the justice of their own point of view and sometimes they vehemently disagree.\n\n\nSuch unexpected dissent led them to clarify why they thpught as they did, often identifying aspects of their personal experience which bad altered their opinions or specific occasions which had made them re-think their point of view. Had the data been collected by interviews the researcber might have been faced with ‘arm chair’ theorizing about the causes of such difference but in a focus group these can be explored ‘in situ’ with the help of the research participants.\n\n\nClose attention to the ways in which research participants tell stories to one another aiso prevents the researcber from assuming that she knows ‘the meaning’ of any particular anecdote or account.\n\n\nParker and Tritter (2006) discuss “key issues relating to the complexity and necessity of considering sampling issues within the context of focus group research and the implications this has for the collection and analysis of resultant data.”\nIdentifies some common logistical impetus for focus groups, relating to acquisition of funding:\n\nIncreasing pressure from research funding organizations to adopt multiple-method research strategies and the fact that focus groups generate far more data than a range of other methods in relation to face-to-face contact between researchers and participants, has added to [the method’s popularity].\n\nCalls out the conflation between focus groups and group interviews:\n\nA similarly pervasive trend in and around academic discussion of qualitative research methods is that focus groups are sometimes seen as synonymous with group interviews and it is this issue which constitutes our second point of contention. … In keeping with the views of a number of other writers in this field, we are of the opinion that there is a fundamental difference between these two research techniques and that the critical point of distinction surrounds the role of the researcher and her/ his relationship to the researched (Smithson, 2000). In group interviews the researcher adopts an ‘investigative’ role: asking questions, controlling the dynamics of group discussion, often engaging in dialogue with specific participants. This is premised on the mechanics of one-to-one, qualitative, in-depth interviews being replicated on a broader (collective) scale. A relatively straightforward scenario ensues: the researcher asks questions, the respondents relay their ‘answers’ back to the researcher. In focus groups the dynamics are different. Here, the researcher plays the role of ‘facilitator’ or ‘moderator’; that is, facilitator/moderator of group discussion between participants, not between her/himself and the participants. Hence, where focus groups are concerned, the researcher takes a peripheral, rather than a centre-stage role for the simple reason that it is the inter-relational dynamics of the participants that are important, not the relationship between researcher and researched (see Kitzinger, 1994a; Johnson, 1996).\n\n\nPart of the problem of achieving this kind of interactional synergy in data collection is that, despite their collective interests, participants may not always be keen to engage with each other, or alternatively, may know each other so well that interaction is based on patterns of social relations that have little to do with the research intent of the focus group. The need to consider the impact on interaction of the constitution of the focus group requires that close attention be paid to methods (and outcomes) of recruit ment. … issues of sampling and selection are likely to prove crucial in relation to the form and quality of interaction in a focus group and therefore the kinds of data one gathers and the extent to which participants share their opinions, attitudes and life experiences.\n\nA large part of what follows is an extremely honest account of the challenges involved in recruiting participants to sit for focus groups. The pressure on students to recruit their friends impacted their views on what views were represented and not represented. Moreover, they draw attention to the role of timing and convenience in recruitment strategy, and representativeness of expressed viewpoints.\nI found it hard to take notes on this since it comes off as a continuous stream of conciousness. However it should still be read from top to bottom to get a sense of practical challenges that are not commonly addressed, save for among veterans and in private circles.\n\nMorgan and Hoffman (2018)" + "objectID": "notes/maelstrom-readings.html#bergeron2018", + "href": "notes/maelstrom-readings.html#bergeron2018", + "title": "Maelstrom reading notes", + "section": "Bergeron et al. (2018)", + "text": "Bergeron et al. (2018)\nThe authors reference the drive for efficiency as a motivating factor that drives open data:\n\nHowever, many cohort databases remain under-exploited. To address this issue and speed up discovery, it is essential to offer timely access to cohort data and samples.\n\nHowever the paper is actually about the need for better and more publicly accessible documentation about data.\nThe authors state that catalogues exist to promote discoverability of data and samples and to answer the data documentation needs of individual studies.\nThey draw attention to the importance of catalogues in research networks (analyzing data across studies), which establish portals that document “summary statistics on study subjects, such as the number of participants presenting specific characteristics (e.g. diseases or exposures)”.\nThe authors outline several challenges that inhibit or limit the potential value of catalogues:\n\nThe quality of a catalogue directly depends on the quality and comprehensiveness of the study-specific information documented. But, maintaining and providing access to understandable and comprehensive documentation to external users can be challenging for cohort investigators, and require resources not always available, particularly for the very small or long-established studies. In addition, the technical work required to build and maintain a catalogue is particularly demanding. For example, gathering comprehensive and comparable information on study designs necessitates the implementation of rigorous procedures and working in close collaboration with study investigators. Manual classification of variables is also a long and a tedious process prone to human error. Moreover, the information collected needs to be regularly revised to update metadata with new data collections. These challenges, among others, can lead to the creation of catalogues with partial or disparate information across studies, documenting limited subsets of variables (e.g. only information collected at baseline) or including only studies with data dictionaries available in a specific language or format.\n\nBullet point summary:\n\nA catalogue’s quality depends on the quality and comprehensiveness of documentation provided by individual studies\nCohort investigators, i.e. leaders of individual studies, are under-equipped to provide such comprehensive documentation\n\nDo they just need material support? Or also guidance on how to do it, what factors to account for, etc?\n\nTechnical work for building and maintaining a catalogue is demanding\n\nI’m not sure if they example they provide to illustrate these challenges aligns with what I would call “technical work”; they refer to precise and detailed documentation in direct consultation with individual study maintainers, and I suppose the discussion about and documentation of methodological details is technical in that it corresponds with the work that was actually done on the ground, using data collection and processing instruments\n\nClassification of variables is a long and tedious process\n\nWhat makes it long and tedious? This isn’t really specified\nThey recite that this is prone to human error, but I wonder what successful or error-ful (?) outcomes would look like and how they would differ\n\nThe information needs to be regularly revised and updated\n\nThe authors recommendations to resolve these concerns:\n\nHowever, to truly optimize usage of available data and leverage scientific discovery, implementation of high quality metadata catalogues is essential. It is thus important to establish rigorous standard operating procedures when developing a catalogue, obtain sufficient financial support to implement and maintain it over time, and where possible, ensure compatibility with other existing catalogues.\n\nBullet point summary:\n\nestablish rigorous standard operating procedures when developing a catalogue\nobtain sufficient financial support to implement and maintain it over time\nwhere possible, ensure compatibility with other existing catalogues" }, { - "objectID": "notes/methodology-notes.html#coding", - "href": "notes/methodology-notes.html#coding", - "title": "Methodology notes", - "section": "Coding", - "text": "Coding\nThese notes are largely derived from my reading of Saldaña (2016), provides a practical overview of what coding entails and specific methods and techniques.\nCoding as component of knowledge construction:\n\nCoding is an intermediate step, “the”critical link” between data collection and their explanation or meaning” (Charmaz (2001), as quoted in Saldaña (2016): 4)\n“coding is usually a mixture of data [summation] and data complication … breaking the data apart in analytically relevant ways in order to ead toward further questions about the data” (Coffey and Atkinson (1996): 29-31, as quoted and edited by Saldaña (2016): 9)\n\nThis relates to the paired notions of decodng when we reflect on a passage to decipher its core meaning, and encoding when we determine its appropriate code and label it (Saldaña 2016: 5).\n\nCoding “generates the bones of your analysis. … [I]ntegration will assemble those bones into a working skeleton” (Charmaz (2014): 113, quoted in Saldaña (2016): 9)\nTo codify is to arrange things in a systematic order, to make something part of a system or classification, to categorize\n\nWhat I sometimes refer to as arranging the code tree\nWhat Saldaña (2016) refers to as categories, I tend to refer to as stubs\n\nCategories are arranged into themes or concepts, which in turn lead to assertions or theories\n\n\nPre-coding techniques\nSaldaña (2016) identifies several techniques for formatting the data to make them easier to code, but also to imbue meaning in the text.\n\nData layout\n\nSeparation between lines or paragraphs may hold significant meaning\nPutting interviewer words in square brackets or capital letters\n\nSemantic markup\n\nBold, italics, underline, highlight\nMeant to identify “codable moments” worthy of attention (Boyatzis (1998), as referenced in Saldaña (2016): 20)\nRelates to Saldaña (2016): 22’s prompt: “what strikes you?”\n\nPreliminary jottings\n\nTri-column exercise with the text on the left, first impression or preliminary code in the middle, and code on the right, after Liamputtong and Ezzy (2005): 270-273.\n\nAsking questions back to the interviewer, or participating in an imagined dialogue\n\nI imagine this might be useful in situations where the time to hold an interview is quite limited and I have to work with limited responses that don’t touch on everything I want to cover\nThe form of questions maintains my tentativity, my unwillingness to commit or assume their responses, and opens the door for their own responses in rebuttal\n\n\nFollowing Emerson et al. (2011: 177), Saldaña (2016: 22) identifies a few key questions to keep in mind while coding:\n\nWhat are people doing? What are they trying to accomplish?\nHow, exactly, do they do this? What specific means and/or strategies do they use?\nHow do members talk about, characterize, and understand what is going on?\nWhat assumptions are they making?\nWhat do I see going on here?\nWhat did I learn from these notes?\nWhy did I include them?\nHow is what is going on here similar to, or different from, other incidents or events recorded elsewhere in the fieldnotes?\nWhat is the broader import or significance of this incident or event? What is it a case of?\n\n\n\n\n\n\n\nNote\n\n\n\nCreate a qc issue for sub-documents, for identifying sections of a document that are especially relevant and hiding less relevant sections. I don’t necessarily want to delete these lines, but I may want to hide them from view. Maybe this is possible using vscode, outside of qc (see https://stackoverflow.com/a/72954133).\n\n\nFrom Clarke (2003: 558) on process coding:\n\nIn a traditional grounded theory study, the key or basic social process is typically articulated in gerund form connoting ongoing action at an abstract level. Around this basic process are then constellated the particular and distinctive conditions, strategies, actions, and practices engaged in by human and nonhuman actors in volved with/in the process and their consequences. For example, subprocesses of disciplining the scientific study of reproduction include formalizing a scientific disci pline, gleaning fiscal support for research, producing contraceptives and other techno scientific products, and handling any social controversies the science provokes (such as cloning and stem cell research)." + "objectID": "notes/maelstrom-readings.html#bergeron2021", + "href": "notes/maelstrom-readings.html#bergeron2021", + "title": "Maelstrom reading notes", + "section": "Bergeron et al. (2021)", + "text": "Bergeron et al. (2021)\nIdentifies several registries of relevant cohorts, but notes that they face challenges getting the data together. Namely, issues concerning institutional policies concerning data-sharing, lack of open access to cohort data and to documentation about the data, the data’s complexity which makes it difficult to harmonize across studies, and lack of access to funding, secure data environments, and specialized expertise and resources.\nThe Research Advancement through Cohort Cataloguing and Harmonization (ReACH) initiative was establihed in collaboration with Maelstrom to overcome some of these barriers in the context of Developmental Origins of Health and Disease (DOHaD) research.\nThe authors briefly summarize some projects that rely on ReACH data, and provide a more comprehensive table of ongoing and unpublished work.\nIn the supplementary materials, the authors also include an illustrative example specific tasks, decisisions and actions that one might got through when using ReACH data. It is a broad-level but fairly sober account of how one would navigate the catalogue and engage with collaborators." }, { - "objectID": "notes/methodology-notes.html#memos", - "href": "notes/methodology-notes.html#memos", - "title": "Methodology notes", - "section": "Memos", - "text": "Memos\nCharmaz (2014: 162) dedicates Chapter 7 to memo-writing, which she frames as “the pivotal intermediate step between data collection and writing drafts of papers.” She locates the power of memo-writing as the prompt to analyze the data and codes early in the research process, which requires the researcher to pause and reflect.\nCharmaz (2014: 162):\n\nMemos catch your thoughts, capture the comparisons and connections you make, and crystallize questions and directions for you to pursue.\n\nCharmaz (2014: 162):\n\nMemo-writing creates an interactive space for conversing with yourself about your data, codes, ideas, and hunches. Questions arise. New ideas occur to you during the act of writing. Your standpoints and assumptions can become visible. You will make discoveries about your data, emerging categories, the developing frame of your analysis — and perhaps about yourself.\n\nCharmaz (2014: 164):\n\nMemo-writing encourages you to stop, focus, take your codes and data apart, compare them, and define links between them. Stop and catch meanings and actions. Get them down on paper and into your computer files.\n\nCharmaz (2014: 164):\n\nMemos are your path to theory constriction. They chronicle what you grappled with and learned along the way.\n\nCharmaz (2014: 165-168) distinguishes between memo-writing and journaling. The former is meant to be more analytical, whereas the latter is more of an account of a direct experience, including significant memories or recollections of moments that stood out (and reflection on why they stood out).\nCharmaz (2014: 171) indicates that “[n]o single mechanical procedure defines a useful memo. Do what is possible with the material you have.” She then lists a few possible approaches to memo-writing:\n\nDefine each code or category by its analytic properties\nSpell out and detail processes subsumed by the codes or categories\nMake comparisons between data and data, data and codes, codes and codes, codes and categories, categories and categories\nBring raw data into the memo\nProvide sufficient empirical evidence to support your definitions of the category and analytic claims about it\nOffer conjectures to check in the field setting(s)\nSort and order codes and categories\nIdentify gaps in the analysis\nInterrogate a code or category by asking questions of it.\n\nCharmaz (2014: 171) draws special attention on bringing data into the memo as a way to more concretely “ground” the abstract analysis in the data and lay the foundation for making claims about them:\n\nIncluding verbatim material from different sources permits you to make precise comparisons right in the memo. These comparisons enable you to define patterns in the empirical world. Thus, memo-writing moves your work beyond individual cases.\n\nThrough a detailed example over the prior several pages, Charmaz (2014: 178) reflects on how memos may “[hint] at how sensitizing concepts, long left silent, may murmur during coding and analysus”. She recalls how writing a memo encouraged her to look back at ideas presented in pivotal texts that she had read earlier in her career, and thereby committed her to a new strand of thought.\nCharmaz (2014: 180) describes how she developed memos from in vivo codes that recurred throughout across the cases. She asked how the saying was applied in different contexts, its overlapping and varied meaning.\nCharmaz (2014: 183-?) encourages adoption of various writing strategies. She notes that “memo-writing requires us to tolerate ambiguity”, which is inherent in the “discovery phase” of research, in which she considers memo-writing to be a part. She encourages adotion of clustering and freewriting techniques to help get the ball rolling (she refers to these as “pre-writing techniques”).\nSaldaña (2016) dedicates Chapter 2 to “writing analytic memos”. Saldaña (2016: 44) notes that codes are nothing more than labels until they are analyzed, and remarks that memo-writing is a stage in the process of getting beyond the data. He refers to Stake (1995: 19), who mused that “Good research is not about good methods as much as it is about good thinking”, and in keeping with Charmaz’s (2014) account of memo-writing, memos are tools for doing good thinking.\nSaldaña (2016: 44-45) channels Charmaz in saying that all memos are analytic memos. I identify with his discomfort in writing according to a pre-defined category of memos, and his preference to categorizing them after writing the memo. He also suggests writing a title or brief description to help with sorting.\nHowever, like Charmaz, Saldaña does differentiate between memos and “field notes”, which are synonymous with Charmaz’s journal entries. According to Saldaña (2016: 45), Field notes are the researcher’s personal and subjective responses to social actions encountered during data collection.\nSaldaña (2016: 53-54) reports on O’Connor’s (2007: 8) conceptualization of contemplation of qualitative data as refraction.\n\nThis perspective acknowledes the mirrored reality and the researcher’s lens as dimpled and broken, obscured in places, operating as a concave or at other times a convex lens. As such, it throws unexpected and distorted images back. It does not imitate what looks into the mirror but deliberately highlights some things and obscures others. It is deliciously … unpredictable in terms of what might be revealed and what might remain hidden.\n\nOther analogies include that by Stern (2007: 119): If data are the building blocks of the developing theory, memos are the mortar”, and by Birks and Mills (2015: 40) who consider memos as the “lubricant” of the analytic machine, and “a series of snapshots that chronicle your study experience”.\nSee Montgomery and Bailey (2007) and McGrath (2021) for more on the distinction between memos and field notes, including detailed examples of these kinds of writing in action." + "objectID": "notes/maelstrom-readings.html#wey2024", + "href": "notes/maelstrom-readings.html#wey2024", + "title": "Maelstrom reading notes", + "section": "Wey and Fortier (2024)", + "text": "Wey and Fortier (2024)\nAn overview of the harmonization procedires applied in CanPath and MINDMAP. Authored by two members of the Maelstrom team, but no one from these initiatives.\nAt first, comes across as another broad-level overview of processes. But, as is elicited in the conslusion, the paper highlights some subtle divergent approaches, some of which are relevant to my project and I picked up here.\nInteresting bit about use of centralized systems or systems that are more conducive to end-users’ individual workflows.\n\nIn both illustrative projects, information about participating studies and data collected was gathered and made available on the central data server. In the CanPath BL-HRFQ, harmonization documentation and processing scripts were held centrally and only updated by Maelstrom Research. In MINDMAP, because multiple groups simultaneously worked on generating harmonization processing scripts for different areas of information, working versions of the DataSchema and processing scripts were held on a GitHub, allowing for better version control and dynamic updating of scripts by multiple remote data harmonizers.\n\nMore about the balance being struck at MINDMAP between institutional and popular tech and documentation platforms:\n\nIn the MINDMAP project, R markdown documents with applied harmonization scripts (Figure 13.1b) and any comments were preserved in the GitHub repository, which is also publicly accessible. Furthermore, summary statistics for MINDMAP variables are only available to approved data users through the secure server, as harmonized datasets are only intended for use within the MINDMAP network. Overviews of the harmonization process and outputs have also been published as open-access, peer-reviewed articles for both projects (Fortier et al. 2019; Wey et al. 2021).\n\nBit about existing collaborative ties making things much easier:\n\nthe prospective coordination and unu sually high ability for the harmonization team (Maelstrom Research) to work directly with study data managers on recently collected and documented study data resulted in high standardization and ability to resolve questions.\n\nData curation as an explicitly creative task, involving impactful decisions. Interesting that this was unexpected enough to warrant inclusion as a key challenge:\n\nIn MINDMAP, these differences were clearly documented in comments about harmonization for each source dataset to allow researchers to decide how to use the harmonized variable. In-depth exploration of the best statistical methods to harmonize these types of measures to maintain integrity of content while minimizing loss of information and methodological bias are important and active areas of research (Griffith et al. 2016; Van den Heuvel and Griffith 2016). The approach taken in this case was to apply simpler harmonization meth ods (i.e. rescaling) and document the transformation, leaving investigators the flexibility to further explore and analyze the harmonized datasets as appropriate for their research questions.\n\nAs with the above excerpt, documentation was considered as a viable mode of resolving or accounting for significant discrepancies:\n\nThe papers describing the harmonization projects attempt to highlight these considerations, for example, providing some comparison of demographics of the harmonized populations against the general populations from which they were drawn (Dummer et al. 2018; Fortier et al. 2019; Wey et al. 2021), listing sources of study-specific heterogeneity in the harmonized datasets to consider (Fortier et al. 2019), and pointing users to individual study documentation where more information on weights to use for analysis should be con sidered (e.g. Wey et al. 2021).\n\nThe bit about use of github was rationalized as a way of facilitating collaboration across insitutional boundaries. They refer to R and RMarkdown being open standards as the main reason, but I wonder if a part of it is that GitHub, being a non-institutional platform, was easier to use from an oboarding perspective:\n\nIn MINDMAP, due to the need for multiple international groups to work simultaneously and flexibly on the harmonization processing and frequently evolving versions of study-specific harmonization scripts, scripts for harmonization processing were written and applied entirely through R markdown in an RStudio interface, and the DataSchema and R markdown versions were maintained and frequently updated in a GitHub repository.\n\nPerhaps ironically, the private company and platform may have been used due to the strength of pan-european collaborative ties that institutions may not be able to keep up with. Whereas in Canada, with centralized project-oriented work, it may be much easier to enforce adoption of centralized tooling. This is just speculation." }, { - "objectID": "notes/methodology-notes.html#preliminary-analyses", - "href": "notes/methodology-notes.html#preliminary-analyses", - "title": "Methodology notes", - "section": "Preliminary analyses", - "text": "Preliminary analyses\nYin (2014: 135-136 5) identifies various strategies for analyzing case study evidence.\n\nA helpful starting point is to “play” with your data. You are searching for patterns, insights, or concepts that seem promising. (Yin 2014: 135)\n\nCiting Miles and Huberman (1994), Yin (2014) lists a few strategies at this playful stage:\n\nJuxtaposing data from different interviews\nPutting information into different arrays\nMaking a matrix of categories and placing the evidence within them\nTabulating the frequency of different events\nPutting information in chronological order or using some other temporal scheme\n\nYin (2014: 135) also emphasizes memo-writing as a core strategy at this stage, citing Corbin and Strauss (2014). These memos should include hints, clues and suggestions that simply put into writing any preliminary interpretation, essentially conceptualizing your data. He uses the specific example of shower thoughts." + "objectID": "notes/maelstrom-readings.html#doiron2013a", + "href": "notes/maelstrom-readings.html#doiron2013a", + "title": "Maelstrom reading notes", + "section": "Doiron, Raina, and Fortier (2013)", + "text": "Doiron, Raina, and Fortier (2013)\nThis paper summarizes what was discussed at a workshop bringing together stakeholders who would contribute to two large data harmonization initiatives: the Canadian Longitudinal Study on Aging (CLSA) and the Canadian Partnership for Tomorrow Project (CPTP). It is therefore representative of plans and challenges that were held at an early stage when collaborations were being established.\nThe authors identify series of reasons for linking data, which I summarize here:\n\nMaximizing potential of disparate information resources\n\n\nenriching study datasets with additional data not being collected directly from study par ticipants\noffer vital information on health outcomes of participants\nvalidate self-reported information\n\n\nDrawing maximum value from data produced from public expenditure\n\n\noffers a cost-effective means to maximize the use of existing publicly funded data collections\n\n\nDevelop interdisciplinary collaborative networks\n\n\nby combining a wide range of risk factors, disease endpoints, and relevant socio-economic and biological measurements at a population level, linkage lays the groundwork for multidisciplinary health-research initiatives, which allow the exploration of new hypotheses not foreseeable using independent datasets\n\n\nEstablish long-lasting infrastructure and instill a collaborative culture\n\n\nLast, a coordinated pan-Canadian cohort-to-administrative linked database would establish legacy research infrastructures that will better equip the next generation of researchers across the country\n\nThe authors use the term “data linkage”:\n\nData linkage is “the bringing together from two or more different sources, data that relates to the same individual, family, place or event”. When linking data at the individual level, a common identifier (or a combination of identifiers) such as a personal health number, date of birth, place of residence, or sex, is used to combine data related to the same person but found in separate databases. Data linkage has been used in a number of research fields but is an especially valuable tool for health research given the large amount of relevant information collected by institutions such as governments, hospitals, clinics, health authorities, and research groups that can then be matched to data collected directly from consenting individuals participating in health research.\n\nThis is distinct from harmonization in that it is not meant to combine data with similar scope and schematic structure, but rather to relate information collected under various domains so that they could be more easily queried in tandem. I imagine this as reminiscient of establishing links between tables in a relational database.\nThe authors identify the open-endedness of the linked data as a unique challenge, without elaborating on this point:\n\nCLSA/CPTP-to-AHD linkage also poses unique challenges in that, in contrast to more traditional requests to link data to answer one-off research questions, it aims to establish a rich data repository that will allow investigators to answer a multitude of research questions over time.\n\nThe workshop participants established a 5-point plan:\n\nbuild strong collaborative relationships between stakeholders involved in data sharing (e.g., researchers, data custodians, and privacy commissioners);\nidentify an entity which could provide overall leadership as well as individual “champions” within each province;\nfind adequate and long-term resources and funding;\nclarify data linkage and data-sharing models and develop a common framework within which the data linkage process takes place; and\ndevelop a pilot project making use of a limited number of linked variables from participating provinces\n\nThe second point, about identifying “champions”, is kind of interesting, and I’d like to know more about what qualities these people were expcted to have, their domains of expertise, their collaborative/soft or technical skills, and how this plays into access to funds and general governance structures\nNeed to look at Roos, Menec, and Currie (2004), which they cite in the conclusion, specifically with reference to the aspiration to develop “information rich environments”. Seems like it is a primary source for the background on linked data in Manitoba and Australia." }, { - "objectID": "notes/methodology-notes.html#the-constant-comparative-method", - "href": "notes/methodology-notes.html#the-constant-comparative-method", - "title": "Methodology notes", - "section": "The constant comparative method", - "text": "The constant comparative method\nThe constant comparative method is based on action codes, similar to what Saldaña (2016) refers to as process codes. According to Charmaz (2000: 515):\n\nThe constant comparative method of grounded theory means (a) comparing different people (such as their views, situations, actions, accounts, and experiences), (b) comparing data from the same individuals with themselves at different points in time, (c) comparing incident with incident, (d) comparing data with categori, and (e) comparing categories with other categories.\n\nMy initial impression is that this is very well suited for Stake’s (2006) multicase study framework, specifically with regards to his notion of the case-quintain dilemma. It also seems very well suited for analysis of situational meaning-making, as per Suchman (1987), Lave and Wenger (1991), Knorr Cetina (2001) and symbolic interactionism at large." + "objectID": "notes/maelstrom-readings.html#fortier2023", + "href": "notes/maelstrom-readings.html#fortier2023", + "title": "Maelstrom reading notes", + "section": "Fortier et al. (2023)", + "text": "Fortier et al. (2023)\nRelates harmonization to the FAIR principles, which has not really been featured much in arguments for harmonization in prior Maelstrom papers. Specifically, this paper frames harmonization as a necessary additional condition that enables FAIR data to be made useful; FAIR is deemed not enough.\n\nIn the following paper, we aim to provide an overview of the logistics and key ele ments to be considered from the inception to the end of collabo rative epidemiologic projects requiring harmonizing existing data.\n\nInteresting acronym/framework for defining research questions:\n\nThe research questions addressed and research plan proposed by harmonization initiatives need to be Feasible, Interesting, Novel, Ethical, and Relevant (FINER). (Cummings, Browner, and Hulley 2013)\n\nTable 1 lists examples of questions that could be addressed to help delineate analytical approach, practical requirements, and operations of a harmonization initiative. These are all questions that look toward or project specific goals, which then inform the strategies through which they may be achieved.\nThe supplementary materials include very detailed information about specific practices and objectives pertaining to the REACH initiative. However, it’s unclear how well this reflects any specific challenges experienced. In other words, I was hoping for something more candid.\nMoreover, this paper is also based on survey responses from 20 harmonization initiatives, but neither the findings resulting from the analysis, nor the data, are referenced or included. Is this the same as the one that informed Fortier et al. (2017)?\nLooming in the background of this paper is the DCC lifecycle model. However they do not cite DCC, or the field of digital curation in general. The DCC lifecycle model has always been presented as a guideline, sort of divorced from practical experience, or at least that’s how I’ve always understood it. Basically, and literally, a model for expected behaviours and interstitial outcomes. I think it would be interesting to explore (a) why the authors perceived need to present a model and (b) how they arrived at the phases and principles that are included in it. Is this tacit or common-sense thinking? Or was this directly informed by my concrete thinking from the field of digital curation?\nI should brush up on more recent work regarding the DCC, and specifically, critiques of it. Through a quick search, a few papers seem directly relevant:\n\nCox and Tam (2018)\nChoudhury, Huang, and Palmer (2020)\nRhee (2024)\n\nI think Cox and Tam (2018) may be especially relevant as a critique of the lifecycle metephor in a general sense. From the abstract, it seems like they identify various downsides pertaining to lifecycle models, specifically that they “mask various aspects of the complexity of research, constructing it as highly purposive, serial, uni-directional and occurring in a somewhat closed system.” I’m not sure how explicit the connection is, but I sense this ties into the general paradigm shift toward greater recognition of curation “in the wild”, as per Dallas (2016)." }, { - "objectID": "notes/methodology-notes.html#situational-analysis", - "href": "notes/methodology-notes.html#situational-analysis", - "title": "Methodology notes", - "section": "Situational analysis", - "text": "Situational analysis\nSituational analysis originates from Strauss’s social worlds/arenas/negotiations framework. From Clarke (2003: 554):\n\nBuilding on and extending Strauss’s work, situational analyses offer three main cartographic approaches:\n\nsituational maps that lay out the major human, nonhuman, discursive, and other elements in the research situation of concern and provoke analyses of relations among them;\nsocial worlds/arenas maps that lay out the collective actors, key nonhuman elements, and the arena(s) of commitment within which they are engaged in ongoing negotiations, or mesolevel interpretations of the situation; and\npositional maps that lay out the major positions taken, and not taken, in the data vis-à-vis particular discursive axes of variation and difference, con cern, and controversy surrounding complicated issues in the situation.\n\n\nClarke (2003: 560) identifies the main purpose of situational data as a way of “opening up” the data, figuing out where and how to enter:\n\nAlthough they may do so, a major and perhaps the major use for them is “opening up” the data –— interrogating them in fresh ways. As researchers, we constantly confront the problem of “where and how to enter.” Doing situational analyses offers new paths into the full array of data sources and lays out in various ways what you have to date. These approaches should be considered analytic exercises — constituting an ongoing research “work out” of sorts—well into the research trajectory. Their most important outcome is provoking the researcher to analyze more deeply.\n\nShe emphasizes that this is meant to stimulate thinking, and should always be paired with comprehensive memoing before, during and after situation mapping excercises.\nA key feature that I think is invaluable is the ability to uncover the sites of silence, or the things that I as a researcher suspect are there but are not readily visible in my evidence. Situational analysis is useful for drawing out the thousand pound gorillas in the room that no one wants to talk about, and is therefore important for identifying things to address during continual data collection, as is one of the (often ignored) central pillars of grounded theory:\n\nThe fourth and last caveat is perhaps the most radical. As trained scholars in our varied fields, usually with some theoretical background, we may also suspect that certain things may be going on that have not yet explicitly appeared in our data. As ethically accountable researchers, I believe we need to attempt to articulate what we see as the sites of silence in our data. What seems present but unarticulated? What thousand-pound gorillas are sitting around in our situations of concern that nobody has bothered to mention as yet (Zerubavel 2002)? Why not? How might we pursue these sites of silence and ask about the gorillas without putting words in the mouths of our participants? These are very important directions for theoretical sampling. That is, the usefulness of the approaches elucidated here consists partly in helping the researcher think systematically through the design of research, espe cially decisions regarding future data to collect.\n\nThe process is remarkably similar to the brainstorming excercise I did with Costis one time. Starts by articulating the actors involved, their roles and relationships, the things they do, the things they make.\n\nThe goal here is to lay out as best one can all the human and nonhuman elements in the situation of concern of the research broadly conceived. In the Meadian sense, the questions are: Who and what are in this situation? Who and what matters in this situation? What elements “make a difference” in this situation?\n\nAfter jotting these down on a canvas or whiteboard, arrange them into more concrete categories. In each category, refer to examples or specific instances. Concepts can occur in multiple categories. By arranging these concepts, grouping them thematically, spatially, and through relationships and associations with arrows or lines (presumably, with labels indicating the nature of these relationships), this provides the researcher with a viable way of unlocking new pathways to think through their data. I imagine this will be especially helpful when arranging or re-arranging the coding tree.\nClarke (2003: 569-570) also dedicates a few paragraphs to relational forms of analysis using this technique:\n\nRelations among the various elements are key. You might not think to ask about certain relations, but if you do what I think of as quick and dirty relational analyses with the situational map, they can be revealing. The procedure here is to take each element in turn and think about it in relation to each other element on the map. One does this by circling one element and mentally or literally drawing lines, one at a time, between it and every other ele ment on the map and specifying the nature of the relationship by describing the nature of that line.\n\n\nYou could highlight (in blue perhaps) that organization’s perspectives on all the other actors to see which actors are attended to and which are not, as well as the actual contents of the organization’s discourses on its “others.” Silences can thus be made to speak.\n\nClarke (2003: 569) On how to generate memos from situational analysis:\n\nEach map in turn should lead to a memo about the relations diagrammed. At early stages of analysis, such memos should be partial and tentative, full of questions to be answered about the nature and range of particular sets of social relations, rather than being answers in and of themselves.\n\nClarke (2003: 570) addresses the energy or mood required to do this kind of work, which is reminiscient of Yin’s (2014: 73-74) description of the “mental and emotional exhaustion at the end of each fieldwork day, due to the depletion of ‘analytic energy’ associated with being attention on your toes.” Interesting that the mention of “freshness” is brought up in the context of Glaser’s tabla rasa approach to grounded theory.\n\nAs a practical matter, doing the situational map and then the relational analyses it organizes can be tiring and/or anxiety producing and/or elating. Work on it until you feel stale and then take a break. This is not the same order of work as entering bibliographic citations. The fresher you are, the more you can usually see. Glaser (1978: 18-35) cautions against prematurely discussing emergent ideas –— that we might not necessarily benefit from talking about everything right away but rather from reflection — and memoing. I strongly agree, especially about early even if quick and dirty memoing. But we all must find our own ways of working best. For most, the work of this map occurs over time and through multiple efforts and multiple memos.\n\nClarke (2003) refers to Shim (2000) as an exemplary case of situational analysis in action." + "objectID": "notes/maelstrom-readings.html#gaye2014", + "href": "notes/maelstrom-readings.html#gaye2014", + "title": "Maelstrom reading notes", + "section": "Gaye et al. (2014)", + "text": "Gaye et al. (2014)\nIntroduces DataShield.\nFrames DataShield as a technical fix to administrative problems:\n\nMany technical and policy measures can be enacted to render data sharing more secure from a governance per spective and less likely to result in loss of intellectual prop erty. For example, data owners might restrict data release to aggregate statistics alone, or may limit the number of variables that individual researchers might access for speci fied purposes. Alternatively, secure analysis centres, such as the ESRC Secure Data Service and SAIL represent major informatics infrastructures that can provide a safe haven for remote or local analysis/linkage of data from selected sources while preventing researchers from down loading the original data themselves. However, to comple ment pre-existing solutions to the important challenges now faced, the DataSHIELD consortium has developed a flexible new way to comprehensively analyse individual level data collected across several studies or sources while keeping the original data strictly secure. As a technology, DataSHIELD uses distributed computing and parallelized analysis to enable full joint analysis of individual-level data from several sources, e.g. research projects or health or administrative data—without the need for those data to move, or even be seen, outside the study where they usually reside. Crucially, because it does not require underpin ning by a major informatics infrastructure and because it is based on non-commercial open source software, it is both locally implementable and very cost effective.\n\nAdds a social/collaborative element to earlier arguments about the challenges inherent of prospective harmonization, highlighting a need for engagement with individual studies (either through direct or peripheral participation) to conduct research that was not initially planned for:\n\nUnfortunately, both [study-level metadata] SLMA and [individual-level metadata] ILMA present significant problems Because SLMA com bines analytical results (e.g. means, odds ratios, regression coefficients) produced ahead of time by the contributing studies, it can be very inflexible: only the pre-planned analyses undertaken by all the studies can be converted into joint results across all studies combined. Any additional analyses must be requested post hoc. This hinders exploratory analysis for example the investigation of sub-groups, or interactions between key variables.\n\nProvides a detailed overview of how DataShield was implemented for HOP (Healthy Obesity Project), including the code used to generate specific figures and analyses. Hoever it does not really describe or reflect upon the processes through which the code was developed.\nThe authors highlight the fact that certain analytical approaches are not possible using DataShield, especially analysis that visualize individual data points. It’s unclear how they enforce this, or whether it’s an implicit limitation based on the data that DataShield participants provide.\n\nBecause in DataSHIELD potentially disclosive com mands are not allowed, some analyses that are possible in standard R are not enabled. In essence, there are two classes of limitation on potential DataSHIELD functional ity: (i) absolute limitations which require an analysis that can only be undertaken by enabling one of the functional ities (e.g. visualizing individual data points) that is explicitly blocked as a fundamental element of the DataSHIELD philosophy. For example, this would be the case for a standard scatter plot. Such limitations can never be circumvented and so alternatives (e.g. contour and heat map plots) are enabled which convey similar information but without disclosing individual data points; (ii) current limitations which are functions or models that we believe are implementable but we have not, as yet, under taken or completed the development work required. As examples, these latter include generalized linear mixed model (including multi-level modelling) and Cox regression.\n\nThe authors list numerous other limitations and challenges. Some have to do with what kinds of data DataShield can handle (something about horizontal and vertical that I do not yet fully understand). Other challenges include the need for data to be harmonized, and having to deal with governance concerns.\nNotably, the first challenge mentioned seems to contradict the statement earlier on (and made by Doiron et al. (2013)) that this is relatively easy to set up. The authors acknowledge the fact that coding for analysis using DataShield has a steep learning curve and requires some pre-planning to enable results from satellite computers to be properly combined. Their mitigation is to black-box these concerns by implementing simpler client-side functions that mask the more complex behaviours (and presumably translate error messages in ways that users can understand and act to resolve!).\n\nDespite its potential utility, implementation of DataSHIELD involves significant challenges. First, although set-up is fundamentally straightforward, application involves a relatively steep learning curve because the command structure is complex: it demands specification of the analysis to be undertaken, the studies to use and how to combine the results. In mitigation, most complex serverside functions are now called using simpler client-side functions and we are working on a menu-driven implementation.\n\nAlso interesting that they note how there may be unanticipated problems, either accidental or malicious, and their way of mitigating against this is to log all commands:\n\nFifth, despite the care taken to set up DataSHIELD so that it works properly and is non-disclosive, it is possible that unanticipated prob lems (accidental or malicious) may arise. In order to iden tify, describe and rectify any errors or loopholes that emerge and in order to identify deliberate miscreants, all commands issued on the client server and enacted on each data server are permanently logged.\n\nThis is even more interesting in light of their continuous reference to “data.care”, which they do not address in depth, but which seems to have been a scandal involving unauthorized release of personal health data used in research.\nThe authors add an additional caveat concerning the need to ensure that the data are cleaned in advance.\n\nBut, to be pragmatic, many of the routinely collected healthcare and administra tive databases will have to undergo substantial evolution before their quality and consistency are such that they can directly be used in high-quality research without exten sive preparatory work. By its very nature, such preparation—which typically includes data cleaning and data harmonization—cannot usually be undertaken in DataSHIELD, because it involves investigating discrepan cies and/or extreme results in individual data subjects: the precise functionality that DataSHIELD is designed to block. Such work must therefore be undertaken ahead of time by the data generators themselves—and this is de manding of time, resources and expertise that — at present — many administrative data providers may well be unwilling and/or unable to provide. That said, if the widespread us ability of such data is viewed as being of high priority, the required resources could be forthcoming.\n\nThis corresponds with another limitation identified earlier, namely with regards to identifying duplicate individual records across jurisdictional boundaries (which involves assumptions regarding nationality and identify – one of those weird myths that programmers can’t seem to let go!):\n\nSo far DataSHIELD has been applied in settings where individual participants in different studies are from different countries or from different regions so it is unlikely that any one person will appear in more than one source. However, going forward, that cannot al ways be assumed. We have therefore been consider ing approaches to identify and correct this problem based on probabilistic record linkage. In the genetic setting 48 the BioPIN provides an alternative solution. Ongoing work is required.\n\nNote the last line of the prior block quote regarding data cleaning:\n\nThat said, if the widespread us ability of such data is viewed as being of high priority, the required resources could be forthcoming.\n\nThis seems like a thread worth tugging at!" }, { - "objectID": "notes/methodology-notes.html#various-other-analytical-strategies-and-techniques", - "href": "notes/methodology-notes.html#various-other-analytical-strategies-and-techniques", - "title": "Methodology notes", - "section": "Various other analytical strategies and techniques", - "text": "Various other analytical strategies and techniques\nYin (2014: 136-142) describes four general analytical strategies:\n\nRelying on theoretical propositions\nWorking your data from the “ground up”\nDeveloping a case description\nExamining plausible rival explanations\n\nYin (2014: 142-168) then goes on to describe five analytical techniques:7\n\nPattern matching\nExplanation building\nTime-series analysis\nLogic models\nCross-case synthesis\n\nRyan and Bernard (2000) describe various analysis techniques for analyzing textual elicitations in structured and codified ways." + "objectID": "notes/maelstrom-readings.html#wolfson2010", + "href": "notes/maelstrom-readings.html#wolfson2010", + "title": "Maelstrom reading notes", + "section": "Wolfson et al. (2010)", + "text": "Wolfson et al. (2010)\nx" }, { - "objectID": "notes/methodology-notes.html#qda-software-and-tooling", - "href": "notes/methodology-notes.html#qda-software-and-tooling", - "title": "Methodology notes", - "section": "QDA software and tooling", - "text": "QDA software and tooling\nWeitzman (2000) provides an overview of software and qualitative research, including a minihistory up to the year 2000 when the chapter was published.\nDescribing the first programs specifically designed for analysis of qualitative data, Weitzman (2000: 804) writes:\n\nEarly programs like QUALOG and the first versions of NUDIST reflected the state of computing at that time. Researchers typically accomplished the coding of texts (tagging chunks of texts with labels — codes — that indicate the conceptual categories the researcher wants to sort them into) by typing in line numbers and code names at a command prompt, and there was little or no facility for memoing or other annotation or markup of text.8 In comparison with marking up text with coloured pencils, this felt awkward to many researchers. And computer support for the analysis of video or audio data was at best a fantasy.\n\nThis history if followed by a sober account of what software can and can not do in qualitative research, as well as affirmation and dismissed of hopes and fears. Very reminiscient of Huggett (2018)." + "objectID": "notes/potential-cases.html", + "href": "notes/potential-cases.html", + "title": "Potential cases", + "section": "", + "text": "Isabel Fortier came up with a shortlist based on consultations to help determine which Maelstrom partner projects may serve as potential cases. We then met on 2025-02-04, when, among other topics, we discussed the shortlist.\nSee the case selection protocol for further details on the parameters that guide how cases are to be determined." }, { - "objectID": "notes/methodology-notes.html#footnotes", - "href": "notes/methodology-notes.html#footnotes", - "title": "Methodology notes", - "section": "Footnotes", - "text": "Footnotes\n\n\nThe term refers to a medieval jousting target: see https://en.wikipedia.org/wiki/Quintain_(jousting)↩︎\nThough Yin (2014: 40-444) is dismissive of such use of the term “sample” since he sees case study research as only generalizable to similar situations, and not to a general population from which a sample is typically said to be drawn. I agree with this focus on concrete situations over Stake’s prioritization of theory-building as an end unto itself.↩︎\nCharmaz uses the term “giving voice” in this specific context. I’m not sure if this is meant to represent Strauss and Corbin’s attitude, and whether this is an accurate representation on their views, but in my mind this should be framed as elevating, amplifying or re-articulating respondents’ voices (and this is a tenet of constructivist grounded theory in general, which derives from Charmaz). My take diverges from the position that we “give voice” to respondents in that it acknowledges (1) that the voices are already there, (2) that respondents are in fact giving us their voices, and (3) that the researcher plays an active editorial role, transforming the respondents’ elicitations into a format that is more amenable to analysis.↩︎\nVery much in line with the pragmatist turn of the late ’90s and early ’00s, as also documented by Lucas (2019: 54-57) in the context of archaeological theory, vis-a-vis positivism, postmodernism, and settling on a middle ground between them.↩︎\nThis seems reminiscient of procedures planned for WP1 and WP2 from my FWO grant application.↩︎\nThis is also very reminiscient of Nicolini (2009).↩︎\nI wonder: would Abbott (2004) call these heuristics?↩︎\nThis caught my eye since its the same approach as that adopted by qc!↩︎" + "objectID": "notes/potential-cases.html#general-notes", + "href": "notes/potential-cases.html#general-notes", + "title": "Potential cases", + "section": "", + "text": "Isabel Fortier came up with a shortlist based on consultations to help determine which Maelstrom partner projects may serve as potential cases. We then met on 2025-02-04, when, among other topics, we discussed the shortlist.\nSee the case selection protocol for further details on the parameters that guide how cases are to be determined." }, { - "objectID": "posts.html", - "href": "posts.html", - "title": "Blog", + "objectID": "notes/potential-cases.html#possible-candidates", + "href": "notes/potential-cases.html#possible-candidates", + "title": "Potential cases", + "section": "Possible candidates", + "text": "Possible candidates\n\nCITF\n\n\nReACH\nTitle: Stress and Anxiety During Pregnancy and Length of Gestation Harmonization Initiative + ReACH\nContact: Julie Bergeron \nReason: A small project and a very complex infrastructure-oriented network coordinated by Julie Bergeron. Both projects are finalized.\nNotes: Julie Bergeron was a PhD student at Maelstrom, and Isabel says that she is probably the most knowledgeable person regarding data harmonization who I will encounter during my research. She worked on her dissertation project (Stress and Anxiety During Pregnancy and Length of Gestation Harmonization Initiative) while also leading work on ReACH, and her dissertation essentially served as one of a few pilot projects under the aegis of ReACH. ReACH was led by Isabel as its PI.\nBoth projects are complete, but Isabel thinks that Julie Bergeron will be able to share some significant insights on this past work. My instinct is that this presents an opportunity to explore how/whether harmonization is suited for doctoral training, the role of pilot projects within broader initiatives, and impact that closeness to the method of data harmonization might have.\nLinks:\n\nhttps://www.maelstrom-research.org/network/reach\n\n\n\nCAPACIty\nTtle: Capacity: Building CAPACIty for pediatric diabetes research and quality improvement across Canada\nAcronym: CAnadian PediAtric diabetes ConsortIum (CAPACIty)\nContact: Shazhan Amed \nReason: Across Canada, focus on clinical data. A lot of work already achieved, and harmonization will start soon. Will use a federated approach for statistical analysis.\nNotes: A network of 15 childhood diabetes centers from across Canada. Went through four years of administrative work, and is now just starting harmonization after finally going through all theose hurldes, despite being very eager to get into the data work early on. Despite these challenges, Isabel thinks they will be very receptive to participating in the study.\n\n\nSHAIRE\nTitle: SHAIRE: Scientific & Health Data Assets In Respiratory Research\nContact: Sanja Stanojevic \nReason: New project just starting, very interesting and dynamic.\nNotes: Extremely new, just got funding very recently. I don’t know that much, to be honest. Could potentially provide redundant value to my study as Capacity, but need to find out more.\n\n\nRespiratory study\nTitle: High-dose docosahexaenoic acid for bronchopulmonary dysplasia severity in very preterm infants: A collaborative individual participant data meta-analysis\nContact: Isabelle Marc \nReason: Very specific and small project, very clinical.\nNotes: A very small project, harmonizing two datasets. I asked if this scale of work is common and Isabel says that it is, so it’s not an outlier.\nLinks:\n\nhttps://maelstrom-research.org/study/n3pi-hi\n\n\n\nMORGAM\nTitle: MOnica Risk, Genetics, Archiving and Monograph\nContact: Kari Kuulasmaa \nReason: European, started several years ago.\nNotes: Older project, ended around 10 years ago, the PI is retired. Might be valuable for looking at broader impact and potential offshoots after the work has been completed.\nLinks:\n\nhttps://www.maelstrom-research.org/network/morgam\n\n\n\nLifecycle\nVery improvized approach to data harmonization, did a lot of “manual” work. According to Isabel, Julie Bergeron will be able to tell me more.\nLinks:\n\nhttps://pmc.ncbi.nlm.nih.gov/articles/PMC7387322" + }, + { + "objectID": "notes/potential-cases.html#recommended-but-no-longer-considered", + "href": "notes/potential-cases.html#recommended-but-no-longer-considered", + "title": "Potential cases", + "section": "Recommended but no longer considered", + "text": "Recommended but no longer considered\n\nCanPath\nTitle: CanPath, formerly called the Canadian Partnership for Tomorrow’s Health\nContact: Noah Frank \nReason: One of the most important harmonization initiative, infrastructure oriented, long-term started more than 10 years ago.\nNotes: Other contacts include John Mclaughlin and Paula Robson.\nMy instinct is to look at how things change over the course of a very long and institutional initiative, especially across discrete phases marked by leadership transitions. But the history here is so vast and I will probably not get much of it through a series of interviews.\n\n\nMindmap\nPromoting mental well-being and healthy ageing in cities. Seems very similar to CanPath in terms of scope and governance, and I would likely face similar challenges." + }, + { + "objectID": "posts/2024-12-09-first-team-meeting.html", + "href": "posts/2024-12-09-first-team-meeting.html", + "title": "Reflection on first team meeting", "section": "", - "text": "Date\n\n\nTitle\n\n\nCategories\n\n\n\n\n\n\nFeb 14, 2025\n\n\nWeek notes (2025-W07)\n\n\nweek notes\n\n\n\n\nFeb 7, 2025\n\n\nWeek notes (2025-W06)\n\n\nweek notes\n\n\n\n\nJan 31, 2025\n\n\nWeek notes (2025-W05)\n\n\nweek notes\n\n\n\n\nJan 25, 2025\n\n\nWeek notes (2025-W04)\n\n\nweek notes\n\n\n\n\nJan 24, 2025\n\n\nOn the role of AI in my research\n\n\nAI / LLM, Methods, QDA\n\n\n\n\nJan 18, 2025\n\n\nWeek notes (2025-W03)\n\n\nweek notes\n\n\n\n\nDec 18, 2024\n\n\nTechnical specs for this website\n\n\nwebsite\n\n\n\n\nDec 9, 2024\n\n\nReflection on first team meeting\n\n\nmeeting notes, general thoughts\n\n\n\n\nDec 9, 2024\n\n\nHello World!\n\n\nintroduction, website\n\n\n\n\n\nNo matching items", - "crumbs": [ - "Blog" - ] + "text": "Last week (2024/12/04) I finally met with David Buckeridge, Tanya Murphy and Aklil Noza in person. The meeting was meant to convey my vision for the project to the whole team, to align perspectives, and to articulate how this will actually work in practice.\nThe gist is that I will be investigating the role of social and cultural factors in data-sharing initiatives such as CITF and other Maelstrom-affiliated projects, and how these relate to, overlap with, or conflict with technical and institutional/administrative factors. To be clear, these are all very important aspects of data-sharing, but we generally recognized that the social and cultural aspects are under-explored relative to their impact.\nWe briefly talked about how we will go about selecting cases, and I emphasized the importance of strategic case selection. This also involves carefully articulating the project’s goals so that the cases will complement them. We agreed that the dataset will likely comprise between 12-15 interviews of around 60-90 minutes in length with representatives from 4-5 cases (one of them being CITF), in addition to representatives of the Maelstrom team. Maelstrom will serve as a “fixed point” that limits the scope of the cases’ breadth and ensures that participants have a common frame of reference. It also potentially allows me to “offload” or “consolidate” reference to technical and administrative aspects of data-sharing through targeted interviews with Maelstrom personnel, instead of dealing with those things with the representatives for each case.\nWe discussed timelines and overlap with Aklil’s work, which will be more concerned with focus groups with CITF databank users. There is definitely overlap with the emphasis of my own work and we will coordinate data collection to enhance the potential for analytical alignment.\nAfter the meeting I chatted with Tanya and Aklil who helped familiarize me with the bigger-picture theoretical discourse and tensions in epidemiology. Much of it seemed familiar since these concerns are common across disciplines, but I still need to read more to concretize my understanding. Tanya recommended I read the “Baby Rothman” which is a condensed version of a very long-lived textbook in this field, among a few other papers she sent me.\nOverall, this meeting got me really excited about this project :)" }, { - "objectID": "posts/2024-12-09-hello-world.html", - "href": "posts/2024-12-09-hello-world.html", - "title": "Hello World!", + "objectID": "posts/2024-12-11-technical-specs.html", + "href": "posts/2024-12-11-technical-specs.html", + "title": "Technical specs for this website", "section": "", - "text": "Welcome to the website for my CITF Postdoc! This will serve as a hub for documenting and sharing my work. I decided to do this as a way of managing and sharing always-updated drafts of research protocols with my supervisor and team members, but it is also generally useful for keeping my thoughts organized. I will also use this blog section to write my thoughts as the project progresses." + "text": "I’m using this website as a way to help organize and share key documents and resources. The research protocols are in flux at this stage in the project’s development, and this will make it easier to distribute up-to-date drafts with partners, while simultaneously enhancing transparency.\nThis post outlines the technical specifications for this website and outlines a roadmap for its further development. It will therefore be continually updated as the site evolves." }, { - "objectID": "posts/2025-01-24-ai-in-my-work.html", - "href": "posts/2025-01-24-ai-in-my-work.html", - "title": "On the role of AI in my research", + "objectID": "posts/2024-12-11-technical-specs.html#fundamentals", + "href": "posts/2024-12-11-technical-specs.html#fundamentals", + "title": "Technical specs for this website", + "section": "Fundamentals", + "text": "Fundamentals\nThis website is based on Quarto, a platform for writing and publishing scientific and technical writing. I had used quarto before but without fully understanding it, and now I am starting to see its elegance.\nI had started off using Hugo, but there were too many limitations that Quarto was able to accomodate. You can find an older version of this post reflecting that setup here: #2346852.\nThe site is hosted on GitHub Pages. The repo is located at https://github.com/zackbatist/CITF-Postdoc." + }, + { + "objectID": "posts/2024-12-11-technical-specs.html#generating-pdfs", + "href": "posts/2024-12-11-technical-specs.html#generating-pdfs", + "title": "Technical specs for this website", + "section": "Generating PDFs", + "text": "Generating PDFs\nAs an avid user, one thing I really like about Quarto is the ability to generate PDFs alongside html versions served over the web. I started tinkering with includes but I need to review how Quarto passes info from YAML frontmatter. It is not at all straightforward and I will need to experiment a bit more with this to get the hang of it." + }, + { + "objectID": "posts/2024-12-11-technical-specs.html#archiving-and-version-control", + "href": "posts/2024-12-11-technical-specs.html#archiving-and-version-control", + "title": "Technical specs for this website", + "section": "Archiving and Version Control", + "text": "Archiving and Version Control\nEvery change is tracked using git. I would also like to archive each research protocol in Zenodo once they reach a point of stability. This would ensure that they ca be aassigned DOIs and detailed metadata, which will make them easier to reference.\nHowever, I do not want to rely on Zenodo’s GitHub integration for two reasons: (1) I want this to be as platform-agnostic as possible, and (2) that system relies on GitHub’s release system which operates on the level of the whole repository rather than specific files.\nI might be able to write a custom CI workflow to archive specific files to Zenodo using their API. But, I want to be able to toggle this option, rather than have it occur for every single detected change. Maybe I can accomplish this by pushing the changes that I want to archive to a dedicated branch that the CI workflow is configured to operate on. Or it might be easier to simply do this manually, since I’m not sure I will be using it that often anyway." + }, + { + "objectID": "posts/weeknotes-2025-W03.html", + "href": "posts/weeknotes-2025-W03.html", + "title": "Week notes (2025-W03)", "section": "", - "text": "AI is upon is, and although I would probably be ok if it wasn’t around, I have been (and still am, to a certain extent) tempted to use it in my research. So here I’m gonna articulate some of my thoughts on AI. This isn’t written to convince anyone, or even to convince myself. Just to lay out all my thoughts and take stock of my preconceptions, disapointments, hopes and desires, etc.\nAlso, I’m gonna use AI, LLM and whatever other brand names and marketing buzzwords interchangably here. Draw whatever conclusions you want about that.\nI see AI as being potentially useful in a few productive activities I regularly engage in:\n\nTranscribing spoken words into written text\nTranscription is a significant component of processing interview data, and this can be extremely slow work. It’s a lot easier to edit a transcript produced through a computer algorithm rather than start from scratch. I used trint, otter and other similar tools before all the AI hype, and more recently I’ve been using whisper to transcribe voice notes that I record while I’m waiting for the bus or drifting off to sleep. I’m not really sure how they’re much different, to be honest. Is AI just a rebrand of natural language processing in these contexts? Either way, I will most certainly be using some automatic transcrion tool in my research.\nSummarizing, breaking down and simplifying complex bundles of ideas\nI do a lot of reading, and it can be hard to get through everything on my list. I therefore make lots of compromises and refrain from reading some things because I just can’t make enough time to get through everything. I imagine that AI can help summarize some key points across a whole corpus of articles on my to-read pile, and I may try it out once I have time to figure out the right tooling for the job. However, I do gain a lot of value from the process of reading. Specifically, as a scholar of scientific practice, I’m interested in the language and rhetoric authors use to describe and situate their methods and findings, and I’m not sure if automatic summary tools can capture and communicate this nuance in ways that I want.\nGenerating code snippets for data processing and visualization\nThis is arguably the most productive potential application I can imagine. Specifically, I’m thinking about using this to generate R code that processes and visualizies data according to imagined outcomes. This is directly relevant to a project I’m working on where I’ve already finished the workflows for scraping and processing the data, I have the questions I want to ask of it, but I don’t have the practical know-how to generate the code that will allow me to address them. ggplot is just so dense to me, and stitching together code snippets from stack exchange is a major pain in the ass that produces a horrible abomination of code that would not pass the muster of any rigorous code review. What’s more, those queries to search stack exchange are already half-formed AI prompts! At least an AI would generate some harmony in the code, and I might learn something by having a tidy and consistent template.\n\nI’m more ambivalent and critical about using AI in these contexts where it’s been really hyped:\n\nAny form of writing, including generating emails and abstracts\nFor me, writing is a creative process and a way of unerstanding. It’s a mechanism through which I come to learn about something. The experience of drafting and revising a document is crucial to my research process. This is especially important for honing my position as a scholar at the intersection of various disciplinary communities, who have distinct language and modes of communication.\nQuerying for truth claims\nTo be clear, the idea that knowledge can be total, absolute and disembodied is deeply flawed, and the popular reception of AI as a neutral observer and reporter of nature makes me sad. That being said, I’m still ambivalent about the potential for specialized, home-grown LLMs as means of parsing, sorting through and obtaining greater value from under-used resources. There are patterns in even the messiest and least formal documents we create, and even if we can’t draw information from these documents, LLMs may be useful to help us reflect on the circumstances of their creation. I keep thinking about Shawn Graham’s twitter bots in this context (which were not based on AI, but whatever), which attempted to spit out segments of artificial reports and fieldwork drama, which real archaeologists often related and resonded to. These responses were interesting to me, often expressed as collective fascination, titilation or disgust, and reminiscient of the apprehension one might experience when hearing your own voice played back while standing at the opposite end of a long hallway. Reacting to distortions of your own experience from very different perspectives can be a really powerful reflexive exercise.\nAs a brainstorming tool, or as a rubber duck\nI’ve heard about people using AI chatbots as agents to bounce their ideas off of. Kind of like eliza, but for productive work. While I think it’s intriguing, I don’t know where I’d start. Also, drawing up the prompt and figuring out how to ask the right questions may already be enough to get the ideas flowing. I think I already do this in some ways by drafting little ephemeral notes, usually directed toward a specific person or imaginary audience while anticipating their feedback. It also somehow seems like a perverse way to de-socialize work, and in a world where students and postdocs feel increasingly isolated, I’d much rather solicit and provide feedback among peers. This has been the foundation of some of my most solid friendships and professional partnerships, and should be encouraged.\n\nI also have some previously-unstated opinions in relation to some common critiques of AI:\n\nProcess versus product\nAI seems to be really good at devising formulaic outputs. That is, it’s good at getting things to look like things whose shapes are already well-defined. This can be valuable in various use cases, like writing emails according to a template or translating texts between languages. I could imagine it being really helpful for those who are coming into a field where certain skills are taken for granted, such as learning how to write “proper” academic emails as a student who is not fluent in english. Imagine being up against a deadline for a job application, while also being knee-deep in unpaid work to get your name out there; an LLM could be a godsend. So I don’t discount easy outputs as inherently bad. A standard output for one is a week-long struggle for another, so I think this distinction between product and process is a false and misleading dichotomy.\nBad instructions\nSometimes I find it really hard to believe that people could earnestly follow whatever an AI tells them. But I think we’re getting to the point of urban mythmaking, similar to the older wariness about following your GPS into a lake. There’s a story behind every warning sign, even if it’s a projection of what you think might happen if you disregard it.\n“Intelligence”\nOne weird thing about AI branding is the smushing together of some unified idea of what constitutes “intelligence”. We’ve already been through this with “smart” gadgets, which have always just been ways to capture consumer products under a platforms proprietary injected plastic molds and information protocols. AI is literally just a way to sell you a new version of the smart gadget you threw out last year.\nTruthiness, i.e., AI’s ability to sound authoritative while also making false claims\nI cringe at any retort to a screenshot of AI giving a wrong definition of a thing. Accuracy of responses should come secondary to critique of the notion that all forms of knowledge can be presented in terms of absolute, disembodied and universally truths. For example, when people ridicule AI’s inability to identify the capitols of various nation states, I see missed opportunities to challenge the value of any answer that anyone might provide. True subversion would be to reject or re-frame the question and the simplicity with which it is addressed.\nOne another related note, I see a lot of weird parallels between myths about truth claims made by AI and by practitioners of qualitative data analysis (QDA) — and, as a qualitative researcher, this is obviously a bit unsettling. Specifically, in both QDA and AI, there is no actual attempt to make absolute truth claims, but the focus is rather on attempting to identify and draw out meaningful elements of elicitations in a corpus, and to trace patterns between them. In my current opinion, the key difference lies in positionality. Any QDA researcher who laim that their cases are representative of all experiences will be laughed out of the room. Meanwhile, AI is lauded for the claims made by their creators that it can derive unambiguous and concrete knowledge from inherently situated and biased data sources. Humility is key while contributing to collective knowledge bases, and AI risks changing the dynamic away from deriving greater value from constructive discourse and toward a system where the loudest voice in the room wins.\nClimate change\nAI uses a lot of energy, and is therefore said to be wasteful. However I think there are certain wasteful components of AI. For instance, generative models that spit out a full sentence to wrap around the answer to a question don’t have to do all that extra work. Also, not everyone is reliant on fossil fuels, and the critique that AI is necessarily bad for the environment is laden with a thick American accent (as is the case with so many of the loudest opinions on the internet).\nThat being said, there are enormous problems with resource allocation in AI, and I’m not trying to dismiss all concerns. I see these concerns as relating to the distribution of power and wealth in society at large, and AI is one aspect of this. Sometimes I wonder if comparisons can be made between using AI in selective research contexts and eating a burger or a banana, which each have their own environmental costs. But thinking in this way is a bit of a trap.\n\nI also see that rhetoric, including anxieties about AI, differs in the various communities I participate in:\n\nIn digital-x, where x = {archaeology | humanities | librarianship | whatever}\nThere’s a lot of experimentation going on. Honestly, I don’t know much about it and I tend to scroll past any discussion about AI applications in archaeology that appears in my feed. Part of me sees it as a passing trend, but it could be better framed as a wild frontier, as is the case with many other things in digital archaeology. People are still in the process of taming the landscape, to make it better suit their needs, and maybe I’ll join in once the settlement is established. But I’m not personally motivated by the dynamism of the current landscape, at least in this particular domain.\nEpidemiology, biostats, public health\nI’m still too new in this community to really make sense of this yet. I’ll continue to watch and learn and listen.\nBroader social science and humanities, as well as libraries, archives and museums\nCritique tends to follow broader, more abstract, and more common-sense lines of thought. In my view, much of this does not really account for the material problems and imperfections in which the social sciences and humanities operate. AI is a lifeline for many people in an overworked, overburdened, under-resourced and hyper-competitive environment, and tut-tutting around how other people use AI sometimes comes across as tone-deaf and disrespectful. Some criticisms of AI being used in real, practical circumstances make me second guess critics’ supposed commitments to improving the social experience of research. The fundamental problem is inequitable access to financial and material resources, and AI’s prevalence is a major symptom of, or — depending on your perspective — resolution to that. People’s who recognize this have no choice but to post broader and more abstract criticisms, which come across as somewhat hollow when disconnected from real and tangible experiences.\nSenior faculty\nProbably the most ambivalent of all communities are senior faculty, who want AI to be useful and will test the waters without fully committing. Which is fine and very prudent, and honestly I identify most with this perspective, despite my position as a lowly postdoc.\nGrad students\nI engage with many grad students. I share my workspace with grad students and encounter them constantly in my day to day neighbourhood forays, where I overhear and sometimes participate in conversations about AI. In my new work environment (Epidemiology, Biostatistics and Occupational Health), the grad students who I engage with have a relatively positive perception of AI. They seem to find greater value in the ability to automate complex processes, using it as a black box of sorts, with predictable and abstracted inputs and outputs, which they see as especially helpful for coding. Outside of this space I’m encountering way more diversity of thought on AI, and I’m not quite sure how to group these viewpoints to structure a proper reaction. I think this in fact contributes to the multitude of perspectives, since no one really cares that much one way or the other to really have a strong opinion (though I sense an overwhelming dissatisfaction when it comes to AI in consumer contexts; this post is largely about productive uses of AI in research and pedagogy).\nI was also told about students learning RStats by just having AI generate their code. The person who pointed this out to me related this to the growing misconception that to learn stats you first need to learn how to code. This in turn relates to the sense that to learn how to do RStats, you just need to memorize a series of steps and copy the text from the slides into the IDE. So, in the end, AI reveals the inadequacy of the teaching mechanisms for programming and stats classes, similarly to how AI has revealed the inadequacy of essay-writing as a pedagogical technique.\nOn the other hand, some students are concerned about dulling their skills, or even not being able to take advantage of opportunities to learn new skills, due to the temptation to automate these tasks. Some upper-year PhD students are glad that they were trained in the fundamentals prior to the AI hype wave. This makes me wonder how students are determining what skills they think they need to know how to do on their own and what is worth running through an LLM. Does it basically operate as a bullshit sensor, where you can smell from a distance that the work is just gonna be tedium and irrelevant? Or is it more out of practical necessity, where you’re stretched so thin that you simply have to rely on these tools to achieve anything meaningful, almost as a mechanism for salvaging one’s work from the claws of austerity? In either case, this points to PhD programs’ inadequacy to match students’ needs and desires, and overwhelming amount of administravia or (seemingly) irrelevant work that students are made to do, which get in the way of their true interests.\n\nMaybe I’ll have more to share some other time." + "text": "I’m trying out a new way to track and communicate my progress on this project. Every week I’ll write a post to track the work I’ve been doing and reflect on my activities. I’ll try to maintain this document throughout the week, tidy it up and post it here on Friday afternoons. However the specific process will probably vary as it grows into the rest of my workflow.\nI’m purposefully trying to not tie this into the personal knowledge management trend. It’s for me and my own purposes, and I don’t want to get bogged down with the unabashed managerial phoniness that belies the PKM phenomenon.\nAnyway, I didn’t take notes on my work this past week, but here’s an overview of what I’ve done from memory:\nContinued to take notes on readings produced by the Maelstrom Project and its partners.\nContinued to investigate and maintain notes on potential cases.\nSet up a placeholder document for notes on methodological concerns.\nMet with David for our bi-weekly check-in (meeting notes are private, at least for now). I was also given access to the lab’s private git server but haven’t really had much of a chance to explore what it’s being used for or devise my own plans to make use of it.\nWorked extensively on the ethics protocol. David and I went back and forth deciding on whether this was necessary, given how the project is based on his grant which already has IRB approval. But it’s better to play it safe than sorry, especially when it’s necessary to obtain informed consent. So to this end, I revised the research protocol and responses to the ethics form, and I also drafted an informed consent document. I’ll share all these things once the whole package is put together (probably in a week or two), but my responses to the ethics form already appears on the ethics protocol page.\nI simplified the way private documents are handled in the quarto project and git respository. May still need to so some fiddling, especially for draft blog posts and notes.\nI started drafting a blog post about the potential use of AI/LLMs in my research. Stay tuned.\nOn a related note, I watched this recent video about the use of LLMs in qualitative data analysis, which did not prompt me to draft the post but which is well-timed, nevertheless.\nI worked a bit more on the data management plan, which prompted me to think more about which QDA software I’ll use. I started filling in a university form to use cloud services provided by MaxQDA, but stumbled upon qualitative-coding (abbreviated as qc), an open source CLI-based QDA system. It represents a very innovative approach to QDA rooted in computational thinking and plain text social science, while also remaining true to the core tenets and purpose of QDA, which make it unique and difficult to design software for. If this is the sort of thing that appeals to you, I highly recommend you read the docs.\nI had a bit of trouble installing it and getting it running, but I met remotely with Chris Proctor, who develops the tool through his work at the Computational Literacies Lab, based in the Department of Learning and Instruction at University at Buffalo (SUNY). He helped me resolve some issues, gave me a guided tour of the system and we just talked about the overall state of qualitative data analysis and its tooling. I don’t really have the capacity right now to post everything he showed me but I will definitely be posting about my experiences tinkering around with qc in the coming weeks.\nSimilarly, I asked on Mastodon about whether there are any tools that might support automatic generation of transcripts that include support for specialized notation. A few linguists and conversation analysis scholars responded with recommendations to use GailBot, and with discussion about the tool’s capabilities and limitations. I requested access to the software but haven’t heard back from the dev team yet. I also created a thread on the whisper github repo, which I now realize it a bit of a naive place to put it, and it hasn’t yet gotten any responses.\nI attended a talk from the epidemiology seminar series, which went wayyyy over my head.\nDid my usual amount of engagement on masotodon, I suppose. And I continued to make new friends in the department too :)" }, { - "objectID": "posts/weeknotes-2025-W04.html", - "href": "posts/weeknotes-2025-W04.html", - "title": "Week notes (2025-W04)", + "objectID": "posts/weeknotes-2025-W05.html", + "href": "posts/weeknotes-2025-W05.html", + "title": "Week notes (2025-W05)", "section": "", - "text": "This week was a bit slower than last. I spent much of it finalizing the content for my IRB application, and the rest preparing for a meeting with a key stakeholder relating to my research.\nThe IRB application is more or less done, just waiting on David and the department head to sign off. It was a major opportunity to re-organize my research protocol and related documents. I shuffled some things over into various placeholder sections in my methodology notes, and pushed a revised research protocol to the website.\nYesterday I posted about on the role of AI in my research. It’s mainly meant to lay out my current state of thinking on AI. I’m not fixed to those ideas, and I think there is much more nuance than I do justice to in that post, but putting it on the page helped me consolidate and put aside some scrambled opinions.\nAfter playing around with qc on Sunday, I started to assemble some feedback. I may post a github issue later this week, once I’ve had a chance to consolidate and edit my thoughts.\nI participated in the weekly CITF logistics update, after which I met with Aklil to discuss the overall strategy for her project and strategize on how we might form the focus groups for that work. We’re gonna meet more regularly, just to share some updates on our respective projects which have a lot in common.\nOn Thursday I met with Isabel Fortier, with the intention of discussing data harmonozation initiatives that might serve as potential cases. It was a bit of a tough meeting. During the first 45 minutes I struggled to communicate the purpose of my work, but I think by the end we reached a greater understanding of what this work will entail and the unique perspective it will bring. One surprising outcome that I still need to think through is Isabel’s suggestion that I slow down a bit, immerse myself more in the world of data harmonization. While she is absolutely right that I’ve been rushing through this first month, I do feel pressure to get the project going and to start accumulating data — I felt a similar pressure when starting my PhD, too. So I need to put my eagerness aside so that the data are relevant and of good enough quality. Isabel offered to schedule regular meetings with me, and even to have me work in the Maelstrom office once a week, and I’m extremely grateful for her support! Plus, I’ll get to have lunch with my mom who works in the same hospital complex, one building over :)" + "text": "The first part of my week largely involved reading and taking notes on methodological texts. I focused on case study design and qualitative coding techniques. I plan to continue my methodology readings on coding techniques, memoing, interview methods and systematic note-taking, as well filling in gaps in my understanding of grounded theory and related debates. These readings are especially useful in this planning stage, but also serve to fill time while I wait for my IRB approval.\nOn that note, I finally submitted my ethics application on Thursday. I expect an expedited review based on the low-risk nature of the work. I posted the materials I submitted on the ethics protocol page.\nI had my biweekly meeting with David, and he was very encouraging.\nI met with Isabel Fortier again yesterday and we came up a list of six projects that may serve as potential cases. We will discuss them in greater depth next week.\nI finally sent some feedback on qc.\nNext week I also need to fulfill a few commitments not as related to the postdoc: I need to work on a peer-review I had committed to, continue assemmbling constructive feedback for qc, and continue going through the DOAJ for the diamond.open-archaeo initiative." }, { - "objectID": "posts/weeknotes-2025-W06.html", - "href": "posts/weeknotes-2025-W06.html", - "title": "Week notes (2025-W06)", + "objectID": "posts/weeknotes-2025-W07.html", + "href": "posts/weeknotes-2025-W07.html", + "title": "Week notes (2025-W07)", "section": "", - "text": "This week was a bit slower than usual. The highlight was my meeting with Isabel to go over details about harmonization procedures and to discuss projects that may serve as potential cases. I got a better sense of the community composition and the kinds of challenges that are commonly experienced, and Isabel was able to provide me with contacts to get in touch with once I’m ready to begin collecting data.\nI was less active with regards to my methodology notes this week, having spent much time re-organizing and consolidating them. I’m also reading about coding techniques and the foundational principles and debates surrounding grounded theory, but there is a lot of ground to cover and I’m taking it all in before I begin recording my thoughts in a systematic way.\nI did some non-postdoc service this week too. I started a peer-review that I had committed to, participated in the monthly SSLA meeting, and completed the first-pass screening of archaeology-related journals in the Directory of Open Access Journals to verify their inclusion in diamond.open-archaeo." + "text": "This week I continued my methodological readings. Today I got some requests for minor modifications to my REB protocol, which will be rather painless to implement.\nI also attended the Public Health day, a student conference where grad students presented their practicum research. I really enjoyed engaging with the students at the poster session, I actually learned quite a lot.\nNext week I’ll try to round off my methodology readings in anticipation of my REB application being approved by the end of this month." }, { "objectID": "qda-protocol.html", diff --git a/sitemap.xml b/sitemap.xml index 2c1cd9b..51418cf 100644 --- a/sitemap.xml +++ b/sitemap.xml @@ -5,88 +5,92 @@ 2025-01-22T21:05:38.131Z - https://zackbatist.info/CITF-Postdoc/posts/weeknotes-2025-W07.html - 2025-02-14T19:53:21.488Z + https://zackbatist.info/CITF-Postdoc/posts/weeknotes-2025-W08.html + 2025-02-21T21:05:33.605Z - https://zackbatist.info/CITF-Postdoc/posts/weeknotes-2025-W05.html - 2025-01-31T18:08:15.277Z + https://zackbatist.info/CITF-Postdoc/posts/weeknotes-2025-W06.html + 2025-02-07T19:44:48.625Z - https://zackbatist.info/CITF-Postdoc/posts/weeknotes-2025-W03.html - 2025-01-18T17:31:25.738Z + https://zackbatist.info/CITF-Postdoc/posts/weeknotes-2025-W04.html + 2025-01-25T16:34:43.500Z - https://zackbatist.info/CITF-Postdoc/posts/2024-12-11-technical-specs.html - 2024-12-18T20:38:01.508Z + https://zackbatist.info/CITF-Postdoc/posts/2025-01-24-ai-in-my-work.html + 2025-01-25T16:06:33.708Z - https://zackbatist.info/CITF-Postdoc/posts/2024-12-09-first-team-meeting.html - 2024-12-18T16:31:38.954Z + https://zackbatist.info/CITF-Postdoc/posts/2024-12-09-hello-world.html + 2024-12-18T16:31:38.955Z - https://zackbatist.info/CITF-Postdoc/notes/potential-cases.html - 2025-02-04T20:47:30.293Z + https://zackbatist.info/CITF-Postdoc/posts.html + 2025-01-19T01:57:01.158Z - https://zackbatist.info/CITF-Postdoc/notes/maelstrom-readings.html - 2025-01-28T20:27:53.768Z + https://zackbatist.info/CITF-Postdoc/notes/methodology-notes.html + 2025-02-21T20:51:37.055Z - https://zackbatist.info/CITF-Postdoc/interview-protocol.html - 2025-01-21T21:35:24.613Z + https://zackbatist.info/CITF-Postdoc/notes.html + 2025-02-19T22:34:54.248Z - https://zackbatist.info/CITF-Postdoc/glossary.html - 2025-01-19T01:58:44.408Z + https://zackbatist.info/CITF-Postdoc/index.html + 2025-01-21T23:24:59.599Z - https://zackbatist.info/CITF-Postdoc/data-management.html - 2025-01-15T16:48:39.050Z + https://zackbatist.info/CITF-Postdoc/ethics-protocol.html + 2025-01-30T21:40:04.298Z + + + https://zackbatist.info/CITF-Postdoc/context.html + 2025-01-31T16:51:55.950Z https://zackbatist.info/CITF-Postdoc/case-selection.html 2025-01-28T15:47:37.243Z - https://zackbatist.info/CITF-Postdoc/context.html - 2025-01-31T16:51:55.950Z + https://zackbatist.info/CITF-Postdoc/data-management.html + 2025-01-15T16:48:39.050Z - https://zackbatist.info/CITF-Postdoc/ethics-protocol.html - 2025-01-30T21:40:04.298Z + https://zackbatist.info/CITF-Postdoc/glossary.html + 2025-01-19T01:58:44.408Z - https://zackbatist.info/CITF-Postdoc/index.html - 2025-01-21T23:24:59.599Z + https://zackbatist.info/CITF-Postdoc/interview-protocol.html + 2025-01-21T21:35:24.613Z - https://zackbatist.info/CITF-Postdoc/notes.html - 2025-02-19T22:34:54.248Z + https://zackbatist.info/CITF-Postdoc/notes/maelstrom-readings.html + 2025-01-28T20:27:53.768Z - https://zackbatist.info/CITF-Postdoc/notes/methodology-notes.html - 2025-02-20T17:31:51.754Z + https://zackbatist.info/CITF-Postdoc/notes/potential-cases.html + 2025-02-21T16:31:42.033Z - https://zackbatist.info/CITF-Postdoc/posts.html - 2025-01-19T01:57:01.158Z + https://zackbatist.info/CITF-Postdoc/posts/2024-12-09-first-team-meeting.html + 2024-12-18T16:31:38.954Z - https://zackbatist.info/CITF-Postdoc/posts/2024-12-09-hello-world.html - 2024-12-18T16:31:38.955Z + https://zackbatist.info/CITF-Postdoc/posts/2024-12-11-technical-specs.html + 2024-12-18T20:38:01.508Z - https://zackbatist.info/CITF-Postdoc/posts/2025-01-24-ai-in-my-work.html - 2025-01-25T16:06:33.708Z + https://zackbatist.info/CITF-Postdoc/posts/weeknotes-2025-W03.html + 2025-01-18T17:31:25.738Z - https://zackbatist.info/CITF-Postdoc/posts/weeknotes-2025-W04.html - 2025-01-25T16:34:43.500Z + https://zackbatist.info/CITF-Postdoc/posts/weeknotes-2025-W05.html + 2025-01-31T18:08:15.277Z - https://zackbatist.info/CITF-Postdoc/posts/weeknotes-2025-W06.html - 2025-02-07T19:44:48.625Z + https://zackbatist.info/CITF-Postdoc/posts/weeknotes-2025-W07.html + 2025-02-14T19:53:21.488Z https://zackbatist.info/CITF-Postdoc/qda-protocol.html