# Tsunami Swami — Frequently Asked Questions At least, what we think you'll ask... ## Positioning & Identity What we are (and aren’t), and where we focus. Q: So… are you an IT company? A consultant? A coach? Or what? A: Not IT support, not generic consulting. We don’t fix your internet or printer problems. We work at the intersection of operations improvement and practical AI enablement: understand how work actually runs, clean up friction, then add automation only if it makes sense. No managed services. No outsourced IT. Just process‑minded problem solving with modern tools. Q: Do you just advise, or do you actually build things? A: Both — but think prototypes, small tools, and systems integrations. If something can be tested quickly, we’ll build it. If it becomes enterprise‑scale, we’ll bring in the right specialists and help structure/manage the work. Quick wins first; scale what proves itself. Q: Do you focus on AI only, or operations too? A: Operations first. AI second. AI only works when the foundation underneath is solid. If there’s chaos, automation just makes chaos happen faster. We stabilize the process before we accelerate it. Q: How are you different from the usual ‘digital transformation’ firms? A: Honestly? I don't know how we're different because I don't really care what the others do. Here's what we do: figure out what's broken, fix the process first, then add tools only if they help. Slide decks, reports, and billable hours aren't the focus — outcomes are. Practical improvements that work for your business, one step at a time. ## Where to Start Entry points based on your confidence and needs. Q: What if we're not sure what AI could do for us — where do we start? A: Pick one: Demystify AI (hands‑on exposure), Operational Deep Dive (understand the current state), or AI Readiness Review (are we even ready?). You don't need a grand plan — just a clear next step. Q: Can we start small before committing to anything big? A: That’s the only way this works. We pick one process, one tool, or one pain point, and test improvement there. If it works, expand. If it stalls, reassess without fallout. Low risk. Clear signal. Quick wins first; then scale what proves itself. Q: Do we need our data or documentation in good shape before we talk? A: No — but we will get there. If documentation is light, we capture how things actually work first. That foundation makes future automation reliable instead of brittle. See Process Documentation. Q: What does a first conversation look like? A: A first conversation is just that — a conversation. You tell us what's going on, what concerns you, what you're curious about. We ask questions, share some initial thoughts, and if it makes sense, talk about where to go next. No pitch, no pressure, no commitment required. ## Delivery & Workflow How we work together and what to expect. Q: Is this workshops and Zoom calls, or hands‑on work? A: Both as needed. We mix short working sessions with hands‑on mapping, cleanup, and small builds. The format flexes; the objective doesn’t: useful outcomes. Q: Do you work remotely or on-site? A: The work dictates the format. A lot of things can be handled remotely, but sometimes being there works better. We'll figure out the best approach based on what has to get done. Q: Do you work with leadership only, or staff too? A: Both — together. Leadership knows intent; staff knows reality. We validate both before making changes. Skipping either creates resistance or fantasy. Q: Do you offer ongoing coaching or support? A: Yes. If a team needs help turning a new idea or tool into habit, ongoing coaching is part of the work — not a separate upsell. Q: Will we get handed off to a junior team once we sign? A: No. You won't be handed off to a rotating cast. You work directly with the same senior operator throughout. If we need extra horsepower — data cleanup, implementation, design — the right people are brought in, but direction and accountability don't change hands. Q: How are engagements structured — project, retainer, or something else? A: Depends on what makes sense. Some engagements have a clear scope — an AI Readiness Review, for example — and run with defined deliverables and an end date. Others are a few focused sessions, like Demystify AI. And some are ongoing by nature, like Swami's Playground. We don't push a model. The structure follows the need. ## Culture, Change & Adoption How we introduce change without chaos. Q: How do you introduce AI without making people feel like they’re being replaced? A: We’re direct: the goal isn’t to eliminate people — it’s to eliminate the parts of their work that drain time or energy. Framed as removing hassle, not humans, adoption turns into curiosity, not resistance. Q: What if our staff isn’t technical? A: That’s common. Demystify AI is designed for non‑technical teams. We focus on plain‑English concepts, hands‑on examples, and using features in tools they already have. Q: How do you make sure new tools or processes actually stick? A: We align incentives, document the new way of working, and support managers to reinforce it. Training + Change Management helps prevent sliding back into old habits. ## Cost, Time & Risk Expectations without making pricing promises. Q: How long do engagements usually take? A: Readiness & Deep Dives: 2–4 weeks. Process Documentation: usually 1–3 workflows at a time. Demystify AI: 3–4 short sessions. Tools & Prototypes: often days or weeks, not months. Goal: visible progress quickly. Q: How much time will you need from our staff? A: Short, focused blocks — typically a few hours per week during discovery and validation, plus async follow‑ups. We protect people’s day jobs by keeping sessions tight and outcomes tangible. Q: Are we talking $5K or $500K? A: We don’t publish pricing, but we sequence work to fit reality. Start small, prove value, then decide together if a larger investment makes sense. If a problem demands enterprise‑scale resources, we’ll structure it accordingly — and only after a clear signal. Q: What happens if we try something and it goes nowhere? A: Then we stop, learn, and adjust. Prototypes are designed to fail cheaply if they’re going to fail. The point is to find signal fast, not defend sunk costs. Q: What happens to our data if AI tools get involved? A: If we're talking about AI and data at all, it's because we've already determined AI can actually help — and that's not always the case. Sometimes a tweak to your current process is all that's needed. If AI is part of the answer, the data needs to be right before it gets near a model. Bad data doesn't just make AI ineffective — it can drive decisions that actively hurt your business. Cleaning up the foundation isn't optional. And the system around the AI has to protect your data. From leaks, from errors, from anything leaving your control without clear agreement. Compliance with industry regulations, legal requirements, and your own internal standards gets built in from the start — not patched on after something goes wrong. ## Credibility & Fit Who this is for, and why trust the process. Q: Have you done this before? A: Operations: 20+ years — large-scale efficiency work, process overhauls, and culture change in public sector operations. Tech: 6+ years — SaaS products and IoT hardware for fleet management. AI: hands-on since late 2022 — I know how it works and where it fails. The consulting practice started in 2022 and builds on all of it. See About for the full story. The short version? I've spent my career being accountable for results, not just recommendations. Q: What kinds of companies is this best for? A: Hands‑on, field‑driven operations (construction, maintenance, logistics, utilities, public works, etc.). Teams that value clarity, speed, and practical results over slide decks. Q: What industries have you worked with? A: Public sector operations (highway maintenance, fleet management, field operations) and tech/software (SaaS, IoT hardware for construction and large contractors). The common thread: operations-intensive work where processes, people, and tools have to align or things break. Q: Do you sign NDAs and handle confidential processes? A: Yes. We work with real operations and sensitive processes; discretion is standard. We'll align on data handling and access up front. Q: Where are you based, and do you work with companies outside your area? A: Indiana. A lot of work — discovery, planning, advisory — can happen remotely and works well that way. If something calls for being on-site, we make that happen too. We figure out what makes sense and go from there. ## Services — Quick Answers Short, practical answers tied to each service. Q: What do you actually look at in an AI Readiness Review? A: Data hygiene, documentation, workflow clarity, tool stack, access/permissions, and where AI could slot in without creating rework. You’ll leave with a prioritized list of fixes and next steps. See AI Readiness Review. Q: Do we get a score or just recommendations? A: A baseline assessment and clear recommendations. If a simple scoring model helps you track progress, we can include that — but the point is actionable steps, not a trophy score. Q: How is an Operational Deep Dive different from a Readiness Review? A: AI Readiness Review checks if foundations are solid for AI. Operational Deep Dive targets how operations actually run to find wasted time, handoff gaps, and process friction. They pair well: tighten operations, then add AI where it helps. Q: What does Process Documentation actually deliver? A: Clear maps of how things actually work (including workarounds), in formats useful to people and machines. That supports training, QA, and future automation. See Process Documentation. Q: Is an AI Roadmap a slide deck or an actual plan? A: A prioritized sequence with owners, prerequisites, and quick wins — not just slides. It clarifies what to do now vs later, and how to avoid overbuilding. See AI Roadmap. Q: Is Demystify AI boring lectures or real examples? A: Short, interactive sessions with real tasks in your existing tools. People learn where AI helps and where it doesn’t. See Demystify AI. Q: What’s the difference between a Tool and a Prototype? A: Tools are quick, lightweight utilities that help now (often AI‑assisted). Prototypes test a process you might scale later, built to prove value in weeks — not months. See Tools & Prototypes. Q: Do you integrate with our existing systems? A: Yes — prototypes and tools are designed to work with what you already have. We use APIs, webhooks, exports, or lightweight connectors. If integration gets complex, we bring in specialists and manage the work. See Tools & Prototypes. Q: Why include Change Management at all? A: New systems fail if people don’t adopt them. Training & Change Management aligns incentives, clarifies impact, and supports managers so the new way sticks. Q: What is Swami's Playground for? A: A safe place for teams to bring ideas and test them without rollout pressure — curiosity with guardrails. Often pairs with Tools & Prototypes. ## Do Now! Things you can do right now. Q: Do we need an AI usage policy even if we're "not using AI" as a company? A: Yes — and it can't be "No AI allowed" unless you want employees to immediately violate the policy and put you further behind. People are already pasting emails, images, or customer details into chatbots or — hopefully for your sake — exploring other AI tools. That's a good thing. But without guardrails, it's also uncontrolled exposure. A simple one-page set of guardrails clarifies what's allowed, what's off-limits, and when to slow down and ask. If you later want a more formal review, that's where an AI Readiness Review comes in. Q: Should we be tracking what AI tools our employees are already using on their own? A: Yes — but lightly. The intent isn't to police curiosity; it's to understand what data is going where and, more importantly, what people are reaching for and why. That's your team telling you where the pain points are and where opportunities exist. A lot of what they're looking for may already be available in software you're already paying for, or there may be a better way to solve it entirely. Helping people spot and use those capabilities is exactly where Demystify AI shines. If real opportunities surface, Tools & Prototypes can help validate them safely. Q: Do we need someone internally responsible for keeping an eye on AI, even before hiring help? A: Not a bad idea. You don't need a formal role and definitely not a committee — just someone who understands tech reasonably well and has a clear line of communication with leadership. Their job is to notice where AI is popping up, surface concerns or opportunities early, and relay anything that might need policy attention. It's a temporary "someone in the know," not a new position. If you later need broader team training, that's where Demystify AI fits. Q: Should we be thinking about how AI might change roles, responsibilities, or risk in the next few years? A: Only if you plan to still be in business in a few years — joking, but also not really. AI is likely going to touch every role, every workflow, and every responsibility in some way, whether you prepare for it or not. You don't need to rewrite org charts, but you do need to look at which tasks AI may assist or automate, where new skills will matter, and where risks could grow. This also applies to hiring and attrition: AI changes which skill sets matter, which roles evolve, and which roles shrink. Thinking about future staffing needs now prevents panicked hiring later. If you want this formalized later, that's an AI Roadmap or Operational Deep Dive conversation, but right now the goal is simple: don't get blindsided. Q: What should leadership be doing right now to avoid falling behind over the next 12–24 months? A: Three things: - Pay attention to real applications, not hype. - Shore up the foundation — outdated tools, scattered docs, permissions nobody understands. If you have experienced people nearing retirement, start capturing how they actually do things before that knowledge disappears. - Encourage small, reversible AI experiments with guardrails. These basics align with the groundwork reviewed in an AI Readiness Review. And yes — hiring Tsunami Swami doesn't hurt either. Q: Do we need guidelines for what data is safe to put into AI tools? A: Absolutely. If someone pastes customer data, HR details, internal pricing, or anything sensitive into AI tools, you've already lost control of it — and depending on your industry, you might also be breaking a law or two. Define what's safe, what's restricted, and who approves exceptions. This eventually folds into broader Process Documentation or readiness work. Q: Should we require a quick human review before AI-generated content goes out the door? A: Yes. AI writes confidently — sometimes incorrectly, and sometimes because it was given the wrong context and didn't know any better. Either way, a human who understands the subject catches what AI can't. There are ways to build systems where not everything needs human review — automated checks, workflow gates, structured processes — and those come later through things like Tools & Prototypes or AI Roadmap work. But if you're reading this, you're not there yet. For now: have a human look at it before it goes out. If your team later needs help reinforcing this habit, Training & Change Management supports that. Q: Our most experienced people are heading toward retirement. How do we keep what they know? A: You won't keep all of it — but you can capture more than you'd expect. The key is starting before someone leaves, not after they've notified you of their looming departure. Document how they actually do things — the real process, not the official one. The workarounds, the judgment calls, the context that never made it into a manual. That becomes onboarding material, training content, and a foundation for smarter tools down the road. This is what Process Documentation is built for. If it's on your radar, don't wait. Q: Should we be budgeting for AI-related costs? A: Yes. AI capabilities are showing up everywhere — as new tools and as additions to software you already use. You're going to use them, and they're going to cost more than what you're paying now. Plan for an increase in spend. If AI is being used effectively, that increase should be more than offset by savings elsewhere — but the budget line needs to exist so it doesn't surprise anyone.