Notes on public failure

"Rhaegar fought valiantly, Rhaegar fought nobly, Rhaegar fought honorably. And Rhaegar died." - George R.R. Martin from A Storm of Swords

That quote has been rattling around in my head all day, because sometimes you do everything right and it still doesn't work out. And when it doesn't work out in front of your entire community, you feel every bit of it.

One of my greatest fears when I decided to apply for Director of Schools in Greeneville was how exposed the whole process would be. Your application is public record. Your interviews happen in front of the board, the staff, the community. People you work with every day are watching. People you go to church with are watching. The parents of the kids in your schools are watching.

There is no quiet rejection letter. No polite phone call. You find out the same way everyone else does, in a room full of people.

I knew this going in. I applied anyway.

When you can't control the outcome, you control what you can. I prepared like my career depended on it, because in some ways it felt like it did. I built data profiles on every school in the district. I studied enrollment trends, assessment data, budget history, staffing patterns. I rehearsed answers until they were sharp and concise. I walked into every round of interviews knowing I had done the work.

And I think it showed. I gave short, direct answers instead of rambling. I spoke from experience, not theory. I was ready.

But preparation doesn't entitle you to anything. It just means you did your part.

In the end, this is a political process. You need 60% of a board to believe you are the right person at the right time. That is three out of five people. You can be qualified, prepared, respected, and still come up one vote short. That is not a reflection of your worth. It is just math.

The person who was selected is a worthy candidate, and I mean that sincerely. I will work hard for him. That is not a line. I have spent my career in service to kids and communities, and that does not change because a vote did not go my way.

People will tell you that failure builds character. I have enough character. What failure actually does is clarify things. It strips away the story you were telling yourself and forces you to look at what is left.

Here is what I know now. I can walk into a room full of people, put myself on the line, and handle the outcome either way. I know that I am good at what I do, not because a board validated it, but because the work I have done over the last two years speaks for itself. And I know that the fear of public failure is worse than the actual thing.

The actual thing just feels like a Tuesday that didn't go your way with a little bit of the five stages of grief sprinkled in.

The Rhaegar quote works because it captures something true: doing everything right does not guarantee the ending you want. Rhaegar was valiant, noble, and honorable. He still lost.

But here is where the analogy breaks down. Rhaegar is dead. I am going to work tomorrow. There are budgets to manage, schools to visit, a salary study to finish, and a hundred other things that matter just as much today as they did last week.

If you are thinking about going for something big and public and risky, do it. Prepare like crazy. Give it your best. And if it does not work out, you will survive. The sun comes up, the work continues, and the people who matter already knew what you were made of before the vote.

"But the wind still blows over Savannah, and in the Spring the turkey buzzard struts and flounces before his hens." - Bukowski, "16-bit Intel 8088 Chip"

Trust Us With the Screens

A couple of weeks ago, I walked through one of our elementary schools during math block. Half the kids had pencils in hand, working through problems on paper. The other half were on devices, running through adaptive lessons that adjusted in real time to what each student actually knew. It was a balanced, intentional classroom. Nobody was doom-scrolling. Nobody was zoned out. And if the Tennessee legislature gets its way with HB 2393, that second group of kids loses their tools.

What the Bill Actually Says

HB 2393/SB 2310, sponsored by Rep. Reneau and Sen. Hensley, originally proposed a near-total ban on digital devices for K-5 instruction. The Senate passed an amended version 31-0 on March 16 that softens the language to require policies governing "age-appropriate and instructional use." But the original House version is still alive, and even the amended version creates compliance burdens and chilling effects on technology use that districts like ours would feel immediately.

The bill carves out exemptions for state-required assessments, IDEA/504 accommodations, and targeted intervention. But the exemptions only highlight the contradiction: if devices are good enough for testing and remediation, why are they suddenly dangerous for instruction?

The Research Is Not What They Think It Is

Much of the legislative momentum traces back to Jared Cooney Horvath's "The Digital Delusion," which synthesizes studies to argue that educational technology performs worse than traditional instruction. It's a compelling narrative. It's also built on shaky ground.

Elizabeth Tipton, a Northwestern statistician who is herself skeptical of ed tech, identified serious methodological problems in Horvath's work. His benchmark comparing technology to "ordinary classroom instruction" is, in her words, "wildly unrealistic." The studies he synthesized vary enormously in quality, and technology evolves so fast that findings from even five years ago may not apply to today's tools.

Here's what the research actually shows: it's complicated.

Recreational screen time is bad for academics. That's consistent across studies. Reading on paper still outperforms reading on screens in most contexts. Phone bans produce modest gains, with larger effects for struggling students. But educational technology, used intentionally, is a different category entirely. The American Academy of Pediatrics updated its guidance in January 2026 to eliminate fixed screen time limits altogether, shifting to a quality-over-quantity framework. Interactive, educational screen time is fundamentally different from passive consumption, and the AAP now says so explicitly.

The honest summary: the evidence calls for balance and intentionality, not a blanket ban. Legislators are reading the research selectively and legislating accordingly.

What We Actually Lose

This is where it gets personal. Greeneville City Schools has been doing things electronically for decades. We're a technology-forward district, and we've built that capacity deliberately. If this legislation passes as originally written, the disruption would be enormous.

Adaptive testing platforms like iReady don't just deliver content on a screen. They use item response theory to pinpoint exactly where a student is, what they've mastered, and what they're ready to learn next. A 2018-19 study found that K-5 students using iReady Instruction with fidelity performed statistically better in reading achievement. These aren't worksheets on a screen. They're diagnostic engines that give teachers actionable data they can't get any other way.

Then there's the testing problem. Tennessee tests students on electronic devices. If K-5 students spend years without touching a device for academic purposes, they'll sit down for state assessments with zero familiarity with the testing environment. That's not a level playing field. That's a setup.

And the cost. We've invested in devices, infrastructure, digital curriculum, and training. Ripping that out and replacing it with physical textbooks and paper-based alternatives would be expensive and wasteful. At a time when districts are already stretched thin, the legislature is essentially asking us to throw away working systems and buy new ones.

Just Let Us Do Our Job

Here's what frustrates me most: we already have the balance right. Walk through our schools and you'll see as many pencils wagging as screens lit up. Our teachers make intentional choices about when technology serves learning and when it doesn't. We don't need Nashville to make those decisions for us.

There's also something worth naming about affect. Students engage differently with adaptive platforms. Some kids who struggle with traditional instruction come alive when they're working through a well-designed digital lesson. That engagement matters, and it's hard to quantify in a legislative hearing.

Sixteen states are now considering similar bills. The political incentive is obvious: screens are an easy villain, and "protect the children" is an unassailable bumper sticker. But good policy requires nuance, and a one-size-fits-all ban is the opposite of nuance.

The Ask

Trust your districts. We're the ones in the buildings every day, watching what works and what doesn't. We've read the research too, and we've done the harder work of actually applying it. If the legislature wants to set guardrails around recreational screen time or social media access, fine. We're already there. But don't take away the instructional tools that are helping our students learn, and then turn around and test them on devices anyway.

I Let a Machine Fill Out My Bracket

Every March, I fill out a bracket with the same strategy: gut feeling, a vague memory of who looked good in February, and the unshakeable belief that this is finally the year a 16-seed makes a deep run. It never works. So this year I tried something different. I built a machine learning model to do it for me, borrowing heavily from the people who do this for real using Claude Code, of course.

Standing on Kaggle's Shoulders

The approach here isn't original. Kaggle runs the March Machine Learning Mania competition every year, challenging data scientists to predict tournament outcomes. I built my model on the foundation laid by Jared Cross's 1st place solution from the 2024 competition, along with the Nate Silver/538 methodology (power rating differential divided by 11, normal CDF) as a sanity check baseline. The competition datasets and the winning approaches are all public, which is what makes a project like this possible for someone who runs a school district by day and trains models by night.

The Data

I pulled 66 datasets covering 18 years of NCAA tournament history. KenPom efficiency ratings, Barttorvik metrics, team resumes, ELO rankings, quad records, shooting splits, coaching histories, conference stats. If someone tracks it, I downloaded it.

From those raw numbers, the model extracts 45+ features for every team: adjusted offensive and defensive efficiency, Dean Oliver's four factors (effective field goal percentage, turnover rate, offensive rebound rate, free throw rate), shooting versatility, defensive pressure, talent ratings, experience, even average height. Then it engineers 10 composite features on top of that, things like quality win percentage and ball security scores.

For every historical tournament game since 2008, the model computes the difference in each of those features between the two teams. That's the training data: 1,070 games where we know who won and by how much.

The Model

A single algorithm wasn't going to cut it. The final model is a stacked ensemble, four base classifiers (logistic regression, gradient boosting, random forest, and histogram gradient boosting) feeding into a meta-learner that weighs their predictions. Think of it as a committee of statisticians who each see the data differently, with a fifth statistician deciding who to trust on any given matchup.

The model hit 73.6% accuracy on a held-out test set. For context, picking the higher seed every game gets you roughly 65%. So the model is finding real signal in the efficiency data beyond what seed lines already tell you.

The single most predictive feature? Barttorvik's adjusted efficiency margin, and it wasn't close. That metric alone carried nearly twice the weight of the next most important feature. Wins Above Bubble, defensive four factors, and assist rate rounded out the top five. Seed difference, the thing most casual bracket-pickers anchor on, mattered less than you'd think.

The Bracket

Once trained, the model does two things. First, it picks a deterministic bracket: for every possible matchup, it takes the team with the higher win probability. Second, it runs 10,000 Monte Carlo simulations, randomly sampling outcomes based on those probabilities, to generate championship odds for every team.

The results:

- Duke: 25%

- Michigan: 18.6%

- Houston: 13.6%

- Arizona: 12.5%

- Purdue: 7.6%

- Illinois: 5.2%

Tennessee, for what it's worth, comes in at 0.2%. I'm choosing to interpret that as "mathematically possible."

The Technical Detour

Here's the part they don't mention in the tutorials. I built this whole thing locally, 1,254 lines of Python, and it wouldn't run. My M3 Max locked up trying to fit the stacked ensemble. This is the second time in two days a machine learning project has tried to kill my laptop (the TN school letter grades analysis did the same thing with XGBoost hyperparameter tuning).

The fix was simple: upload everything to Google Drive and run it in Colab. Free cloud compute, all the sklearn dependencies pre-installed, no kernel panics. The whole pipeline, data loading, training, 10,000 bracket simulations, HTML generation, ran in a couple of minutes.

If you're doing ML work on a Mac and things start freezing, don't fight it. Just move to Colab. Your laptop will thank you.

What I Actually Learned

The model confirmed something I already suspected: efficiency margins are the whole game. Not record, not conference strength, not recruiting rankings. How many points you score per possession versus how many you allow. That's it. Everything else is noise or a downstream effect of that core metric.

It also reminded me that 73.6% accuracy means the model is wrong more than one game in four. March Madness is chaotic by design. Single-elimination tournaments reward variance, and no amount of feature engineering will predict the kid who hits a half-court buzzer-beater.

But that's the fun of it. The model gives you a framework, a set of informed probabilities. What you do with those probabilities is still up to you. The full code, data, and a live tracker comparing predictions to actual results are on https://github.com/jasonbhorne/march-madness-2026.

You can see how it’s doing below.

Predicting Achievement Without Test Scores

I Can Predict Your School's Achievement Without Looking at a Single Test Score

A machine learning analysis of roughly 1,700 Tennessee public schools across two years, comparing what letter grades tell us versus what they hide.

Tennessee gives every public school a letter grade. A through F, just like report cards. The state calculates it from a formula that weighs achievement scores, growth, chronic absenteeism, English learner progress, and for high schools, graduation rates and college/career readiness.

The formula is public. If you know a school's test scores, you can basically calculate the grade yourself. Which raises a question I've been chewing on: what if you strip out all the test-based inputs and just look at the structural stuff, the demographics, staffing, funding, discipline rates, the conditions a school operates under? How much can you predict?

The answer surprised me.

The Experiment

I pulled every publicly available dataset from the Tennessee Department of Education for the 2022-23 and 2023-24 school years: letter grades, school profiles, chronic absenteeism, discipline, educator experience, teacher retention, staffing ratios, per-pupil expenditures, funding sources, graduation rates, and dropout rates. Merged them all at the school level. About 1,690 eligible schools per year, observed across both years for 3,381 school-year observations.

Then I deliberately removed every variable that directly feeds Tennessee's letter grade formula. No achievement scores, no growth scores, no success rates, no CCR rates. What remained were 33 contextual features: things like percent economically disadvantaged, chronic absenteeism, teacher retention, per-pupil spending, and demographic composition.

I ran the analysis two ways. First, I tried to classify the letter grade itself (A through F). Then I switched the target to overall success rate, the continuous achievement percentage that drives the letter grade. Same features, different targets. The comparison is telling.

Round 1: Predicting the Letter Grade

Five models. Random Forest, XGBoost, Gradient Boosting, Logistic Regression, and an Ordinal Logistic model that respects the A > B > C > D > F ordering. Best accuracy across the board: about 40%.

Model Accuracy CV Accuracy Mean Absolute Error
Logistic Regression 41.8% 39.9% 0.73 grades
Ordinal Logistic 41.4% 40.1% 0.74 grades
Random Forest 40.2% 40.3% 0.75 grades
XGBoost 34.4% 40.6% 0.83 grades
Gradient Boosting 37.1% 39.6% 0.79 grades

40% accuracy across five categories is better than random (20%), but not great. The models were off by about 0.75 letter grades on average. If a school is a C, the model might guess B or D. Close, but noisy.

The letter grade bins are doing real damage here. A school with a 49% success rate and a school with a 51% success rate might land in different grade buckets, but structurally they're nearly identical. The model sees the same features and reasonably groups them together, but the grading system draws an arbitrary line between them.

Round 2: Predicting Achievement Directly

Same 33 contextual features. Same schools. But instead of predicting A/B/C/D/F, I targeted the overall success rate, a continuous percentage from 5% to 95%.

Model comparison showing R-squared values for all seven regression models

R-squared comparison across models. Gradient Boosting and XGBoost both explain over 81% of variance in achievement.

Model R-squared Mean Absolute Error CV R-squared
XGBoost (Tuned) 0.823 5.5 pct pts
Gradient Boosting 0.816 5.6 pct pts 0.819
XGBoost 0.815 5.7 pct pts 0.822
Random Forest 0.759 6.4 pct pts 0.783
Ridge Regression 0.698 7.2 pct pts 0.663
Linear Regression 0.698 7.2 pct pts 0.615
Lasso 0.689 7.3 pct pts 0.661
R² = 0.82 Contextual features alone explain 82% of the variance in school achievement. No test scores needed.
±5.5 pts The tuned model predicts a school's success rate within 5.5 percentage points on average.

That is a massive jump. The same features that could only guess a letter grade 40% of the time can explain 82% of the variance in achievement when you let the model see the actual number instead of a bucketed label.

Scatter plot showing actual vs predicted achievement, tightly clustered around the diagonal

Actual vs. predicted achievement. Points cluster around the diagonal, with an MAE of about 5.5 percentage points.

What Drives Achievement

SHAP (SHapley Additive exPlanations) tells us not just which features matter, but how much they move the needle and in which direction. The units here are percentage points of achievement.

SHAP feature importance showing economically disadvantaged percentage and chronic absenteeism dominating

Feature importance measured by mean absolute SHAP value. Two features dominate everything else.

Two features tower over the rest:

  • Economically disadvantaged percentage: 5.3 points of influence on average. Higher poverty, lower achievement.
  • Chronic absenteeism: 4.7 points of influence. More absent students, lower achievement.

After those two, a cluster of second-tier features emerges: local funding percentage (positive), demographic composition, experienced teachers (positive), teacher retention (positive), and discipline rates (negative). Each of these contributes roughly 0.6 to 1.3 percentage points.

SHAP beeswarm plot showing feature effects on achievement predictions

SHAP beeswarm plot. Each dot is one school. Red means high feature value, blue means low. Dots pushed right increase the predicted success rate, dots pushed left decrease it.

Look at that SHAP summary. High economically disadvantaged percentage (red dots) consistently pushes predictions left (lower achievement). High chronic absenteeism does the same. High local funding and experienced teacher percentages push right (higher achievement). The patterns are clear and consistent.

Why the Comparison Matters

The letter grade classification flopped not because the features lack signal, but because the grading system collapses a continuous reality into five bins. A school at the 49th percentile and a school at the 51st percentile might be structurally identical, but one gets a C and the other a B. The model can't distinguish them because there's nothing structurally distinguishing to find.

When you let the model predict the actual achievement percentage, it stops fighting artificial boundaries and starts learning the real relationship between conditions and outcomes. The same data that produced a mediocre 40% classifier produces an R-squared of 0.82 when you ask the right question.

This is a data science lesson wrapped in education policy. If your outcome variable is discretized from something continuous, you're throwing away information. The letter grade system takes a rich, nuanced distribution of achievement and flattens it into a handful of buckets.

Distribution of achievement and achievement by letter grade showing overlap between grade categories

Left: the actual distribution of achievement across Tennessee schools. Right: the same data, grouped by letter grade. Notice the overlap, especially between B, C, and D schools.

A Case Study: Greeneville City Schools

I work for Greeneville City Schools, so I ran our numbers through the same lens. The model says poverty and absenteeism explain 82% of achievement. GCS has a district-wide economically disadvantaged rate around 29%, which puts us in the middle of the pack. Based on structural factors alone, the model would predict us to land somewhere around the state average.

We don't.

+8.1 pts In 2023-24, GCS scored 8.1 percentage points above the expected achievement for districts with our demographic profile, nearly double the 4.2-point gap from the year before.
15th of 98 Among districts with similar ED populations, GCS ranked 15th in achievement in 2023-24, up from 26th the year prior.

In 2023-24, four of our seven schools earned A grades. Here's every GCS school, year over year:

School ED % 2022-23 2023-24 Change
Eastview Elementary 18% 56.9% (A) 61.2% (A) +4.3 pts
Tusculum View Elementary 27% 41.8% (B) 50.0% (A) +8.2 pts
Greeneville High School 24% 50.0% (A) 48.4% (A) -1.6 pts
Greeneville Middle School 24% 44.9% (B) 47.8% (A) +2.9 pts
Hal Henard Elementary 36% 49.7% (B) 48.3% (C) -1.4 pts
Highland Elementary 54% 32.6% (C) 36.5% (C) +3.9 pts
TOPS Greeneville 17% 29.1% (D) 37.5% (C) +8.4 pts

Five of seven schools improved, several significantly. Tusculum View jumped from a B to an A with an 8.2-point gain. TOPS Greeneville climbed 8.4 points and moved from a D to a C. Even Highland Elementary, our highest-poverty school at 54% ED, scored 36.5%, well above the 24% state average for schools in that ED range. Highland ranks 13th out of 131 schools with similar poverty levels statewide.

The model says schools like ours should perform at a certain level given our demographics. We keep outperforming that prediction, and the gap is widening. That's not an accident. That's what happens when experienced teachers stay (we have strong retention), absenteeism is managed, and the district invests in the things that actually move the needle.

What This Means for Districts

If you run a school district in Tennessee, here is what 1,700 schools, two years of data, and seven models are telling you:

  • Your letter grade is 82% predictable from factors that have nothing to do with how well you teach. Poverty and absenteeism alone account for most of the variance.
  • The two highest-leverage things a district can invest in are reducing chronic absenteeism and supporting economically disadvantaged students. Everything else is a rounding error by comparison.
  • Teacher experience and retention matter, but they're second-tier effects. A school with great teachers in a high-poverty, high-absenteeism context will still struggle on paper.
  • Spending more money per pupil, counterintuitively, correlates negatively with achievement. This isn't because money hurts. It's because Title I funding flows to the schools that need it most, and need is correlated with the same factors that drag down scores.

None of this is new to anyone who runs schools. We all know poverty predicts outcomes. But there's a difference between knowing it and seeing a machine learning model explain 82% of the variance with nothing but contextual features. It puts a precise number on something we've felt in our bones for years.

The uncomfortable implication: Tennessee's letter grade system is, to a large degree, grading the ZIP code. A school's structural context is doing most of the talking, and the letter grade is mostly just a noisy echo of it. But districts like Greeneville show it doesn't have to be destiny. The 18% of variance the model can't explain? That's where the work happens.

Methodology Notes

Data: Tennessee Department of Education public data downloads for 2022-23 and 2023-24. All school-level. Schools flagged as ineligible for letter grades were excluded. Approximately 1,690 unique schools observed across both years, yielding 3,381 school-year observations (3,345 with valid achievement data).

Features: 33 contextual variables across demographics, teacher quality, discipline, absenteeism, finance, staffing, graduation, and dropout. All formula-input features (achievement scores, growth scores, success rates, CCR rates) were deliberately excluded.

Models: Seven regression models (Linear, Ridge, Lasso, ElasticNet, Random Forest, Gradient Boosting, XGBoost) plus hyperparameter tuning via RandomizedSearchCV. Five classification models for the letter grade comparison. 80/20 train/test split, stratified. 5-fold cross-validation on training sets.

SHAP values computed via TreeExplainer on the XGBoost regression model. All code available on request.

2024-25 TN School Letter Grades: Third Year Analysis

The Tennessee Department of Education released the 2024-25 school letter grades on December 18, 2024. This is the third year of letter grades under the revamped formula that emphasizes academic achievement over growth. In this post, I continue my analysis of these letter grades, examining distribution trends, demographic correlations, and standout schools.

Data Sources

The data files used for this analysis are available from the Tennessee Department of Education's data downloads page. I merged the 2024-25 Letter Grade File with the 2024-25 School Profile data to examine demographic patterns.

Out of 1,905 schools listed for letter grades, 208 (10.9%) were ineligible to receive a grade. These schools were excluded from this analysis, leaving 1,697 eligible schools.

Distribution

The 2024-25 letter grades were distributed as follows:

  • A: 355 (20.9%)

  • B: 483 (28.5%)

  • C: 491 (28.9%)

  • D: 302 (17.8%)

  • F: 66 (3.9%)

This continues the positive trend we've seen over the past three years. The percentage of A schools has increased from 17.4% in 2022-23 to 20.9% in 2024-25, while F schools have decreased from 5.4% to 3.9%.

Distribution Comparison

Year-Over-Year Comparison

Grade 2022-23 2023-24 2024-25 3-Year Change
A 17.4% 19.0% 20.9% +3.5%
B 26.1% 26.8% 28.5% +2.4%
C 30.4% 29.8% 28.9% -1.5%
D 20.7% 19.6% 17.8% -2.9%
F 5.4% 4.8% 3.9% -1.5%

The data shows steady improvement: more schools are earning A's and B's while fewer are receiving D's and F's.

What Influences a Grade

Average Scores by Letter Grade

Grade Achievement Growth Growth25 Success Rate LG Score
A 4.85 4.94 4.29 58.2% 4.83
B 4.06 3.78 3.47 44.5% 3.91
C 3.21 2.52 2.91 35.4% 2.94
D 2.08 1.61 2.71 23.7% 2.01
F 1.00 1.00 2.90 12.5% 1.18

A notable pattern persists from previous years: schools with an F grade actually show higher Growth25 scores (2.90) than D schools (2.71). This metric measures the progress of the lowest-performing 25% of students. While F schools are making gains with their struggling students, this improvement is not sufficiently weighted to improve their overall letter grade.

Summary Dashboard

Subgroup Analysis

The relationship between demographic factors and letter grades remains significant.

Economically Disadvantaged Students

Grade 2022-23 2023-24 2024-25
A 18.3% 22.3% 20.5%
B 28.9% 31.1% 29.2%
C 36.5% 38.9% 35.0%
D 42.0% 45.9% 41.7%
F 54.5% 49.6% 56.8%

Black, Hispanic, Native American Students

Grade 2022-23 2023-24 2024-25
A 21.8% 35.8% 27.4%
B 27.7% 48.7% 36.3%
C 39.5% 56.0% 42.0%
D 55.0% 65.8% 52.5%
F 80.1% 82.2% 84.5%

The gap between A schools and F schools remains substantial. A schools average 20.5% economically disadvantaged students compared to 56.8% in F schools. For BHN students, the gap is even more pronounced: 27.4% in A schools versus 84.5% in F schools.

Subgroup Three-Year Comparison

Correlation Analysis

The correlations between demographic factors and letter grade scores show a consistent pattern across all three years:

Subgroup 2022-23 2023-24 2024-25
Economically Disadvantaged -0.50 -0.44 -0.47
Black/Hispanic/Native American -0.37 -0.39 -0.34
Students with Disabilities -0.09 -0.13 -0.12

The correlation between economically disadvantaged percentage and letter grade score (r = -0.47) indicates a moderate negative relationship. Schools with higher percentages of economically disadvantaged students tend to receive lower letter grades.

Correlation Scatterplots

Correlation Heatmap

Conclusions

The 2024-25 letter grades show continued improvement across Tennessee schools. Key findings include:

1.    More schools are earning A's (20.9%, up from 17.4% three years ago) and fewer are receiving F's (3.9%, down from 5.4%).

2.    The correlation between poverty and letter grades remains strong (r = -0.47), but some high-poverty schools continue to beat the odds.

3.    F schools continue to show strong Growth25 scores, meaning they are making progress with their lowest-performing students, but this is not reflected in their overall grades.

The letter grade system continues to simplify complex accountability measures into easily digestible grades. While this provides transparency for families and communities, it's important to remember that these grades are heavily influenced by factors outside of schools' direct control, including poverty and segregation. Schools serving high-need populations face steeper challenges in achieving high letter grades, making the success of many schools all the more remarkable.

Also, it is important to note that many schools in Northeast TN (Johnson, Carter, Unicoi, Greene, and Washington Counties) went through major flooding in the fall, and they worked very hard to just have school, much less show improvement in accountability data.

This analysis used Python with pandas, matplotlib, seaborn, and scipy for data processing and visualization. I used Claude to do the coding and proof my writing.

An Open Letter to the TN State Board of Education Re: World Language Credits

Dear Members of the Board:

When I was in college at ETSU, I had the chance to study in Paris twice. We stayed at the Cité Universitaire, a community built after World War I so students from around the world could live and learn together. The purpose was simple: understanding. That, at its core, is what studying a world language gives our students.

Learning another language opens doors. Even the basic knowledge gained in high school helps students communicate, connect, and participate more confidently in a global economy. If our goal is to prepare young people for the future, giving them at least an introduction to another language seems essential.

It also expands their understanding of people beyond our borders. Many of our students may never travel internationally, but world language classes give them access to other cultures, traditions, and ways of thinking. These courses build cultural awareness, empathy, and curiosity: qualities that strengthen citizenship at home. These are qualities we hope to see in the citizens of Tennessee. 

Mark Twain once said that travel is “fatal to prejudice, bigotry, and narrow-mindedness.” The study of world languages offers a similar antidote. It broadens young people’s perspectives and helps them see the world, and others, with clearer eyes.

Please do not eliminate world languages as a graduation credit. Tennessee already provides flexibility for students who wish to pursue an additional program of study in place of World Languages and Fine Arts. There is no compelling reason to remove this requirement entirely, and doing so would limit the opportunities we give our students.

Finally, consider what unintended consequences this would have on World Language programs in small, rural counties. Large metropolitan districts could absorb the cost of this, but schools on smaller budgets would have to consider eliminating these programs altogether. 

Thank you for your consideration.

Jason B. Horne, Ed.D. 

French Teacher
Assistant Director of Schools, Greeneville City Schools

Building an AI-Powered Fantasy Football Draft Assistant with NotebookLM

When researching fantasy football strategies, you’ll notice that they almost all say the same thing: draft Ja’Marr Chase or Justin Jefferson in the first round. Beyond that, it gets a little hazy. Some say to go heavy with WRs and some say to go heavy with RBs, and some say don’t even look at a RB or QB until the 5th round. Sifting through all of this great advice is a slog, and ultimately, you’re left a bit more confused than you were in the beginning. This year, I’m turning to AI to organize this information into something that I can use both ahead of time and in real time to draft my fantasy football team.

My AI strategy

My plan is to use NotebookLM as my draft day guide and the place where all of my research ends up. You can see my notebook at the link above.

In the notebook, I had ChatGPT generate a deep research prompt to use. This is what it came up with:

"Fantasy Football Deep Research Prompt I am preparing for a 10-team PPR fantasy football league with the following setup: Roster Size: 16 players Starters: 9 Bench: 7 (+1 IR) Positions: QB: 1 (max 4) RB: 2 (max 8) WR: 2 (max 8) TE: 1 (max 3) FLEX: 1 (RB/WR/TE) D/ST: 1 (max 3) K: 1 (max 3) Scoring System: Head-to-Head Points, PPR (point per reception). Draft Type: Snake Draft. Research Goals Draft Rankings & Strategy Provide updated player rankings and tiers for QB, RB, WR, TE, FLEX, D/ST, and K in a 10-team PPR league. Identify early-round must-target players at each position based on projected points per game, positional scarcity, and upside. Recommend value picks and sleepers (players likely to outperform ADP). List players to avoid (injury risk, overvalued, declining production). Positional Breakdown Ideal draft strategy by round (e.g., RB-heavy early vs. WR priority). Depth chart analysis: when to prioritize QB vs. TE vs. FLEX depth. Defense & kicker strategy (when to draft, which matchups to target). Schedule & Playoff Outlook Identify players/teams with the most favorable fantasy playoff schedules (Weeks 15–17). Highlight key players with tough matchups to avoid. Advanced Analytics Use expected fantasy points (xFP), target share, snap counts, and red-zone usage to identify breakout players. Incorporate injury history and workload trends into rankings. Compare upside vs. floor for mid-round picks. League-Specific Adjustments How roster depth (7 bench spots, 1 IR) should affect draft strategy. Best balance between high-floor starters vs. high-upside bench stashes. Waiver wire strategy recommendations given a 10-team setup (shallower league, stronger FA pool) This should give you a highly targeted research pack with rankings, projections, draft plans, and league-specific strategy tailored to your exact format.

I then used this same prompt for ChatGPT, Google Gemini, Claude, and Perplexity.

I downloaded each result to a PDF and then added it my NotebookLM. I generated a Mindmap, Podcast, and Presentation from NotebookLM (mostly for fun). The mindmap is very useful.

My plan is to talk to NotebookLM in real-time to let it know who has been drafted to give me a suggestion for my next pick. I will also put the same questions into AI. I only have 60 seconds to draft my next player, so I will have to be fast and be ready to go.

I’ll let you know how it turns out. Last year, I forgot about the draft only 15 minutes after sitting down to do it, and I finished 2nd in the league, so maybe this year will be better (or much worse).

Enjoy this video overview that my NotebookLM generated from my fantasy research docs, and don’t miss the podcast that it generated.

Bluebooks for Bluebloods? Rethinking Writing, Assessment, and AI in Today’s Classrooms

As AI tools like ChatGPT become more common in schools, some educators are reaching for bluebooks in an attempt to restore academic integrity. But are we asking the right questions? In this post, I explore whether writing should remain the default method for assessing knowledge, or if it’s time to reimagine our approach. From speeches to projects to presentations, there may be better ways to engage students in deep, meaningful learning—especially in a world where AI is part of the process, not just a threat to it.

Read More

How to Show Up and Stand Out in an Interview

How to Show Up and Stand Out in an Interview

A student asked me for advice for an assistant principal interview. I rattled off some things, and then I had ChatGPT fact-check them, and it turned into this blog post.

Interviews are more than just a series of questions—they’re a chance to show who you are, how you think, and why you're a great fit for a leadership role in a school. Here’s a research-supported guide that blends practical advice with professionalism.

Be Authentic

People can spot inauthenticity quickly. Trying to be someone you're not is hard to maintain—especially under pressure. Research shows that authenticity in high-stress interviews improves how candidates are perceived (Krumhuber et al., 2022). Be yourself. Be prepared. That’s more than enough.

Make It a Conversation

The strongest interviews feel like a dialogue, not an interrogation. When candidates ask thoughtful questions and stay conversational, it builds rapport and demonstrates interest. SHRM (2021) notes that hiring panels consistently rank these candidates higher.

Be Specific About Why You Want THIS Job

Avoid generalities like “I’m ready for a change” or “I want a promotion.” Talk about why you want this job in this school with these people. Research on person-organization fit shows that alignment with values and mission significantly increases hiring likelihood (Kristof-Brown et al., 2005).

Bring Work Products and Materials

Come with a clean, professional copy of your resume, letter of intent, and any relevant work samples. A 2023 LinkedIn survey found that 74% of hiring managers value candidates who bring a portfolio or artifacts to support their answers.

Have a Laptop with You

You may be asked to show something electronically—lesson plans, data reports, or digital tools. Being able to pull something up shows you're prepared and tech-capable.

Take Notes

Even if it’s just jotting down a few words, taking notes signals engagement and gives you a moment to think before responding. According to Forbes (2021), it also makes you appear focused and professional.

Dress the Part

Your clothing sends a message before you speak. Professional dress still impacts perceived competence and leadership. For men, a jacket tends to increase perceived authority. For women, conservative neckline choices receive more serious consideration—largely due to unconscious bias (Psychology Today, 2022; Bègue et al., 2019). This isn’t about style policing—it’s about managing perception.

Carry Something Grounding

I always bring a bright-orange Yeti to high-stakes meetings. It helps keep me calm. Research supports the idea that small comfort objects can lower anxiety and improve focus (Clinical Psychological Science, 2018).

Scan the Room When You Talk

Use the “thirds rule”: spend part of your time making eye contact with the center, part with the left, and part with the right side of the panel. This creates a sense of inclusion and presence (Toastmasters International).

It’s Okay Not to Know

If you don’t have a perfect answer, say so. Ask a clarifying question, take a breath, and gather your thoughts. Hiring managers respect honesty and thoughtfulness over a shaky bluff.

What Not to Do

  • Don’t lie.
  • Don’t exaggerate.
  • Don’t make things up.

Honesty and humility remain two of the top-rated traits hiring committees look for in leadership roles (Indeed Hiring Lab, 2022).

Final Word

Be humble. Be confident. Be prepared.
But most of all—be yourself.

References

My greatest concern for AI in education: the invisible handshake

I had a mentor teacher who was fairly unethical, and once said to me,
“Horne, if you start getting complaints, give ’em all A’s. Those complaints will dry up.”

It was meant as a joke (I think), but like most bad jokes, it carried too much truth. The idea was simple: avoid scrutiny by keeping everyone happy even if it meant compromising the core purpose of education: teaching and learning.

That memory has been on my mind lately as I think about the role of AI in schools.

What happens when a teacher uses AI to design the lesson… AI to generate the assessment… students use AI to complete it… and the teacher uses AI to grade it… and everyone makes an A?

Nobody complains.
But nobody learns.
And everyone quietly agrees not to say the obvious part out loud: This isn’t real.

That’s my greatest concern—not that AI will destroy education, but that it will dull it. That we’ll settle for the illusion of learning because it’s easier, faster, quieter. That we’ll stop asking, “Did they grow?” and start asking, “Did it look good?”

We could end up with a system where everyone is satisfied—teachers, students, parents, even administrators—and yet nothing of substance is happening. It looks like learning, but it’s just going through the motions.

That’s the invisible handshake.

Tennessee School Letter Grades: A Machine Learning Update

In early 2024, I published a blog post exploring how machine learning could predict Tennessee school letter grades based on demographic data. That analysis provided insight into the structural factors influencing school performance metrics, particularly the significant role of economic disadvantage. Since then, I’ve expanded the dataset to include both the 2023 and 2024 school letter grades, refining the model to capture trends over time and further assess the predictive power of demographic variables. This post serves as an update on that work and a precursor to a more detailed research paper.

Revisiting the Predictive Model

Tennessee assigns letter grades (A–F) to schools based on multiple criteria, including student achievement, academic growth, and college/career readiness. While these grades are intended to reflect school quality, they often correlate strongly with socioeconomic and demographic factors.

For this updated analysis, I merged data from both years to improve model robustness. The methodology remained consistent:

  • Data Collection: Letter grades and demographic data were sourced from the Tennessee Department of Education.

  • Data Processing: Letter grades were converted to a numeric scale, and missing values were cleaned.

  • Machine Learning Models:

    • Random Forest Regression for feature importance analysis

    • Linear Regression to estimate the impact of individual factors

    • Correlation Analysis to identify relationships between variables

  • Letter Grade Scale

    A 4.5-5.0

    B 3.5-4.4

    C 2.5-3.4

    D 1.5-2.4

    F 1.0-1.4

Key Findings

Economic Disadvantage Remains the Strongest Predictor
Schools with higher percentages of economically disadvantaged students continue to receive lower letter grades. The model estimates that for every 10% increase in economically disadvantaged students, the expected letter grade drops by 0.35 on a 1.0–5.0 scale.

Feature Importance Score

Special Education Enrollment Also Impacts Letter Grades
The presence of students with disabilities (SWDs) is another contributing factor, though to a lesser extent. A 10% increase in SWD enrollment is associated with a 0.11-point decrease in letter grade.

Race and Economics Are Closely Linked
Schools with higher percentages of Black, Hispanic, and Native American students tend to have lower grades. However, when controlling for economic disadvantage, racial composition becomes a less significant predictor. This suggests that economic factors, rather than race itself, drive these disparities.

See the correlation matrix below:

Correlation Matrix

Demographics Explain About One-Third of Grade Variation
The updated linear regression model explains 29% of the variance in school letter grades, meaning that while demographics play a measurable role, other factors—such as instructional quality, funding, and school leadership—are also crucial.

What If Demographics Were Removed?
A simulated model scenario where a school had no economically disadvantaged students and no SWDs predicted an average letter grade of 4.5 (A range). This reinforces the idea that many low-performing schools face structural challenges beyond their control.

Public Schools: A Testament to Resilience

In the context of Tennessee’s voucher era, public schools operate under significantly different—and often more challenging—conditions than private schools, making their achievements all the more remarkable. The assumption that private schools inherently provide a superior education doesn’t hold up under scrutiny.

Unlike private institutions, which can be selective in their admissions, public schools serve every student—regardless of economic status, disability, or other challenges. Private schools often have the luxury of limiting enrollment to students who fit their preferred criteria, while public schools are tasked with educating all children who walk through their doors—often with fewer resources and greater accountability.

Despite these challenges, many public schools deliver exceptional outcomes. Educators work tirelessly to support students from diverse backgrounds, implementing innovative instructional strategies and targeted interventions to ensure student success. Their ability to foster academic achievement and social growth, even in the face of structural obstacles, speaks to the public education system's resilience, dedication, and effectiveness.

Public schools should be celebrated rather than maligned for their role in serving all students, strengthening communities, and proving that educational excellence is not exclusive to selective institutions.

This blog post was proofread with suggested corrections by Grammarly.