-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
CalData - Benefits Recommender collaboration next steps #31
Comments
We currently use query parameters in our links to BenefitsCal.com and getCalFresh.org. We don't have direct Google Analytics to either of these accounts. BenefitsCal doesn't use GA (I think they use amplitude instead). These partners provided the query parameters we should use for them to track inbound links. We depend on them sharing stats with us periodically. We receive a spreadsheet with the status of applications started from inbound traffic from BenefitsCal. We haven't received data from getCalFresh.org yet but expect it soon. |
Thanks @aaronhans -- what about the referral links to things that do fall under the ca.gov domain (WIC, water assistance...)? Or do I misunderstand the construction of the links? |
We aren't using any tracking parameters for any other links like WIC, LIHEAP, LIHWAP. We also have noopener & noreferrer set on the offsite a tags so in those cases the partner site shouldn't be able to identify the source of our traffic. @reganking can elaborate but I believe our partners like WIC, LIHEAP, LIWHAP aren't likely able to share data with us. The water programs have different instructions for each county which sometimes end up with a phone number to call instead of a web form. |
So, would it be fair to say that we currently do not have a way to identify traffic coming from the recommender widget on any of the |
Correct Ian. cc @reganking |
This is great. I'm unfamiliar with "CBO-style". How manually-generated are these reports? Are they structured tabular data (e.g., excel, csv)? If so, should we do the legwork to start including them in the data warehouse for our modeling needs? - Ian I shared the BenefitsCal weekly report with you: @ian-r-rose and @britt-allen Since this report provides the most detailed "downstream" referral partner data, it makes sense to view this with respect to supporting an application flow diagram. - Regan
Awesome. We're planning to do take advantage of some rapid onboarding training from our contract with dbt (@britt-allen has been spearheading that), though we may wait until one or two more of our team are hired/onboard (February?). But if you're interested in that I'm sure you'd be welcome. In the meantime, since you already have some familiarity, here is our dbt repo. We can get you a seat in our dbt cloud account, and I also put some effort into documenting a local workflow if you feel like getting a local Python environment set up (I prefer this to the cloud UI). - Ian Thanks Ian ;) I will check it out and would also like to train with the group. - Regan
We're discussing with @kimberlyhicksodi different ways we might engage on experiment design, so I'd imagine that we may have some recommendations on how to set up the experiment schema to be maximally useful. In the meantime, we'd be interested in looking at an extract of the current table if you have it handy. Airtable may be possible (I haven't kicked the tires on how easy it is to jailbreak that data), but it's likely there are easier systems to integrate. - Ian Here's our worksheet for BR Experiments based on the list in AirTable. We created the original list prior to launching the widget and have, thus far, tested #1 (in usertesting.com protocol; results not linked), #20-30, with last test #21 (head-to-head) ending yesterday. These are not all really experiments. I have added the uniform parameters to each experiment in the list for my own assessment needs. We do not currently have: a formal protocol for prioritization experiments, determining the impact of experiment outcomes, or reporting of experiments. I look forward to discussing this with @kimberlyhicksodi - Regan
Can you elaborate on what you mean here? Do you already have models or columns for these, or are you looking forward to what we would like to do? - Ian Only the parameters added in the worksheet, and this is for assessment only at this point. Here's a draft data model and our AirTable Experiment pipeline data definition. The model captures entity relationships I'm interested in with respect to scaling UX multi-variate tests (e.g., language, content, position in a layout, design style, etc.). - Regan
I didn't mean anything too fancy by "self-consistent". Was mostly hoping we could say something like: "The number of GA4 views with XX query parameter is within 5% of the number of widget clicks referring to that page". But I guess that is not possible with the current setup. - Ian That would be great, of course. Capture of gA from referral partners CSD (LIH(W|E)P) and CDPH (WIC) is a goal. Discussions this week hopefully can move this forward a bit. The WIC referral basically points to a ZIP code entry / phone number but with noreferrer set, we're limited. gA data also from search console and trends could help us fill in the gap understanding what the size of our placement page audiences are compared to those searching in California for benefits in general. - Regan
We could try to match based on timestamp. It would be circumstantial, but I'd view this as largely a gut-check at this point. I'm curious whether you or @aaronhans think that would be a useful exercise. - Ian Well, WIC clicks already give us the amount of traffic we're generating. Our discussion with WIC on Friday should help us understand where they see their program going, in terms of gA measurement in general. They've only been operational with a new team within the last month, so this is a really an ideal discussion partner. - Regan
Do we already have this? If not, perhaps we can get CDT to ask them to add the statewide GA tag. - Ian No. We're meeting with Rob McAndrews team to discuss. They are willing. We can discuss adding the statewide. - Regan
I'd love to see early previews of these reports if you have any. - Ian Me too ;) I can try to get an Amplitude report from BenefitsCal. I opted for an excel weekly report. CfA promised us access to their custom dashboard. - Regan
Do you have any decks or other materials on the keyword analysis? I haven't seen it yet. - Ian I can walk you though what I did here in the BR enhanced keyword analysis. - Regan
Very interesting, that would be great to see as well! - Ian Let's look at the BenefitsCal weekly report I shared with you. It's only capturing counties. - Regan |
Closing as complete, pending new activity we may revisit, but nothing is planned right now |
Thanks @ian-r-rose and @britt-allen. We're re-grouping as we leave the benefits pilot. We will select a couple top priority items to look at together. And, give you time to choose what makes sense for CalData. Most of the GA issues are resolved so this should be more interesting to look at in the next round. |
The CalData team had a productive meeting with the CalInnovate benefits recommender team yesterday to discuss their data needs and how to move forward with our collaboration. Here are the notes from that meeting. This morning, as part of CalData DSE's spring planning we sketched out a proposal for next steps over our upcoming sprint (2023-01-09 through 2023-01-20), which this issue intends to capture. The below points could spin out to separate issues as necessary.
1. Investigate the needs around data update frequency
@sarah-letson and @jpetrucione are frequent consumers of the Looker dashboards, and have needs around data freshness and reliability. In particular, @sarah-letson asked if we could increase the data loading frequency from the current six-hour cadence. We can! In order to better serve the needs of the project, I'd like to understand:
2. Investigate needs for different filters and aggregations of the benefits widget data.
@mediajunkie was interested in having more aggregated views of the data easily available. For instance, we might want to break down a Clicks/Views ratio by several different facets:
Such aggregation may be possible but painful in Looker, and doing some pre-modeling in dbt could be a more natural fit.
3. Investigate pain points around analyzing and joining data that exists today
Our understanding of the datasets that exist today are:
It sounds like one pain point is around designing a schema for the experiment variations such that it can be easily joined with widget data to get descriptive statistics of experiment results. Some follow-up things we want to learn:
4. Determine how to leverage the currently-collected Google Analytics data
A longer-term analysis of this project might include tracking of user journeys from benefits site -> recommender -> benefits site -> completed application -> approved application through Google Analytics data. That would likely require some significant planning and engineering in partnership with downstream agencies. However, we did learn about a query parameter put in the URL of the referral links, and thought it might be fruitful to look for that in our data:
The text was updated successfully, but these errors were encountered: