Do You Still Believe in These Common Tech Myths?

Tanya Wetson-Catt • 8 August 2023

In today's digital age, technology plays a significant role in our lives. But along with the rapid advancements and innovations, several myths have persisted. 


Is it okay to leave your smartphone charging overnight? Do Macs get viruses? And what about those 5G towers? What’s going on with those?


Common tech myths can often lead to misunderstandings. They can even hinder your ability to fully use various tools and devices. In this blog post, we will debunk some of the most common tech myths that continue to circulate. We’ll also explore the truth behind them.


Myth 1: Leaving your device plugged in overnight damages the battery


First is one of the most persistent tech myths. Leaving your device plugged in overnight will harm the battery life. But this myth is largely outdated.


Modern smartphones, laptops, and other devices have advanced battery management systems. These systems prevent overcharging.


Once your device reaches its maximum charge capacity, it automatically stops charging. This is true even if it remains connected to the power source. In fact, it is often recommended to keep your device plugged in overnight to ensure a full charge by morning.


So, feel free to charge your gadgets overnight without worrying about battery damage.


Myth 2: Incognito mode ensures complete anonymity.


Many users believe that using incognito mode in web browsers guarantees complete anonymity. They feel completely secure while surfing the internet using this mode. But this is not entirely accurate. While incognito mode does provide some privacy benefits, they’re limited.


For example, it mainly prevents your device from saving the following items:


  • Browsing history
  • Cookies
  • Temporary files


However, it does not hide your activities from your internet service provider (ISP). Nor from the websites you visit. ISPs and websites can still track your IP address. They can also still watch your online behavior and collect data.


Do you truly want to remain anonymous online? Then consider using a virtual private network (VPN). Or other specialized tools that provide enhanced privacy protection.


Myth 3: Macs are immune to viruses.


Another prevalent myth is that Mac computers are impervious to viruses and malware. It is true that Macs have historically been less prone to such threats compared to Windows PCs. This does not make them immune. 


Some people that tout this myth point to malware statistics. For example, in 2022, 54% of all malware infections happened in Windows systems. Just 6.2% of them happened in macOS.


But you also need to factor in operating system (OS) market share. As of January 2023, Windows had about 74% of the desktop OS share. Mac’s OS had just 15%.


When you consider this, it turns out the systems aren’t that different when it comes to virus and malware risk. The infection rate per user on Macs is 0.075. This is slightly higher than on Windows, at 0.074. So, both systems have a pretty even risk of infection. This is the case even though Macs have a significantly lower infection count.


As the popularity of Macs has grown, so has the interest of hackers in targeting these devices. Malicious software specifically designed for Macs does exist. Users should take proper precautions, no matter the operating system in use.


You need to install reliable antivirus software. As well as keeping the operating system and applications up to date. Exercise caution when downloading files or clicking on suspicious links. Being aware of potential security risks and practicing safe browsing habits is crucial. This is true for Mac users, just as it is for any other platform.


Myth 4: More megapixels mean better image quality


When it comes to smartphone cameras, savvy marketing sometimes leads to myths.


Many people believe that more megapixels equal better image quality. This is a common misconception.


Megapixels are an essential factor in determining the resolution of an image. But they are not the sole indicator of image quality. Other factors play a significant role. Such as:


  • The size of individual pixels
  • Lens quality
  • Image processing algorithms
  • Low-light performance


A camera with a higher megapixel count may produce larger images. But it does not guarantee superior clarity, color accuracy, or dynamic range.


Manufacturers often strike a balance between pixel count and other image processing technologies. They do this to achieve optimal results. When choosing a smartphone or any camera, consider the complete camera system. Don't only focus on the megapixel count.


Separate Fact from Fiction


In a world where technology is an integral part of our lives, you must separate fact from fiction. Debunking common tech myths can empower you to make informed decisions. It can also maximize the potential of your digital experiences. An understanding of the truth behind these myths helps you use technology more effectively. It can also help you better protect your privacy.


Get the Technology Facts from a Trusted Pro


Whether you need help with an infected PC or setting up a corporate network, we’re here for you. We cut through the tech myths to bring you reliable and efficient service.


Give us a call today to chat about your technology goals and challenges.

Let's Talk Tech

More from our blog

by Tanya Wetson-Catt 20 March 2026
Since cloud computing became mainstream, promising agility, simplicity, offloaded maintenance, and scalability, the message was clear: “Move everything to the cloud.” But once the initial migration wave settled, the challenges became apparent. Some workloads thrive in the cloud, while others become more complex, slower, or more expensive. The smart strategy for 2026 is a pragmatic hybrid cloud approach. A hybrid cloud strategy blends public cloud services like AWS, Azure, and Google Cloud with private infrastructure, whether that’s a private cloud in a colocation facility or on-premise servers. The goal isn’t to avoid the cloud, it’s to use it wisely. This approach recognises that one size does not fit all. It gives you the flexibility to place each workload where it performs best, considering cost, performance, security, and regulatory requirements. Treating hybrid as a temporary solution is a mistake, as it is increasingly becoming the standard model for resilient operations. The Hidden Costs of a Cloud-Only Strategy Relying on a single model can create blind spots. The cloud’s operational expense (OpEx) model is fantastic for variable workloads . but for predictable, steady-state applications, it can cost more over time than a capital investment (CapEx) in on-premise equipment. Data egress fees, the cost of moving data out of the cloud, can lead to surprise bills and create a form of “lock-in.” Performance can also suffer. Applications that require ultra-low latency or constant, high-bandwidth communication may lag if they’re forced into a cloud data centre far away. A hybrid approach lets you keep latency-sensitive workloads close to home for optimal performance. The Strategic Benefits of a Hybrid Cloud Model First, a hybrid cloud strategy is all about balancing resilience and flexibility. For example, during peak periods like a holiday sales rush, you can take advantage of the public cloud’s scalability and then scale back to your private infrastructure when demand drops. This approach can significantly reduce costs. Second, hybrid cloud helps meet data sovereignty and strict compliance requirements. You can keep sensitive or regulated data on infrastructure you control while running analytics or other workloads in the cloud. This setup is often essential for healthcare, government, finance, and legal sectors, where data must remain within a specific legal jurisdiction. According to FedTech , hybrid cloud gives government agencies the best of both worlds, allowing innovation while meeting strict security standards. Why Some Workloads Need to be kept On-Premise There are several scenarios where private infrastructure makes the most sense: Legacy and proprietary applications: Some organisations run systems that are difficult to move to the cloud, either because of security requirements or simply because they perform better and cost less on-premise. Large-scale data processing: When moving data out of the cloud could trigger high egress fees, it can be more cost-effective to run applications on-site. Predictability and control: Certain workloads require consistent performance and precise control over hardware. Real-time manufacturing systems, high-frequency trading platforms, or core database servers often perform best on dedicated, on-premise infrastructure. Build a Cohesive Hybrid Architecture The main challenge of a hybrid cloud is complexity. You’re managing two or more environments, and success depends on how well they integrate and are managed. That’s why reliable networking is essential, a secure, high-speed connection between your cloud and on-premise systems, often through a dedicated Direct Connect or ExpressRoute link . Unified management is just as important. Use tools that provide a single dashboard to track costs, performance, and security across all environments. Containerisation, using platforms like Kubernetes, can also help by allowing applications packaged in containers to run smoothly in either location. Implement Your Hybrid Strategy Start by auditing your applications and categorising them. Which ones are truly cloud-native and scalable? Which are stable, legacy, or sensitive to latency? Mapping your applications this way will highlight the best candidates for a hybrid approach. Begin with a non-critical, high-impact pilot. A common example is using the cloud for disaster recovery backups of your on-premise servers. This tests your connectivity and management setup without putting core operations at risk. From there, migrate or extend workloads strategically, one at a time. The Path to a Future-Proof IT Architecture Adopting a hybrid mindset creates a future-proof IT architecture. It reduces the risk of vendor lock-in, preserves capital, and provides a built-in safety net. The cloud landscape will keep evolving, and a hybrid foundation lets you adopt new services without a full rip-and-replace. It also allows you to move workloads back on-premise if that makes sense for your business. The goal for 2026 is intelligent placement, not blind migration. Your infrastructure should be as dynamic and strategic as your business plan, and a blended approach gives you the flexibility to make that happen. Reach out today for help mapping your applications and designing the hybrid cloud model that best fits your business goals. Article FAQ Does a hybrid strategy mean I failed at moving to the cloud? Not at all. It means you matured beyond a simplistic “all-in” approach. It demonstrates a sophisticated IT strategy that prioritises business outcomes over technology dogma. Many of the world’s largest tech companies use hybrid models. Is hybrid cloud more secure? It can be. It allows you to apply the most appropriate security model to each workload. You can keep your most sensitive data in a private, air-gapped environment while still leveraging the cloud’s advanced security tools for less-sensitive applications. The key is managing the secure connection between the two. What is the biggest challenge with a hybrid setup? The main challenges lie in the complexity of resource management and networking. With inadequate planning and/or implementation, you can end up creating two isolated silos instead of having a unified environment. As such, invest in skilled architecture and unified management tools to overcome this.
by Tanya Wetson-Catt 16 March 2026
Imagine a former employee, maybe someone who didn’t leave on the best terms. Their login still works, their company email still forwards messages, and they can still access the project management tool, cloud storage, and customer database. This isn’t a hypothetical scenario; it’s a daily reality for many small businesses that treat offboarding as an afterthought. Many businesses don’t realise how much access departing employees still have. When someone leaves, every account, login, and permission they had must be carefully revoked. If offboarding is disorganised, it creates an “insider threat” long after the employee is gone. The risk isn’t always malicious, often, it’s simple oversight. Old accounts can become backdoors for hackers, forgotten SaaS subscriptions continue to drain funds, and sensitive data may remain in personal inboxes. Failing to revoke access systematically is an open invitation for trouble, and the consequences range from embarrassing to catastrophic. The Hidden Dangers of a Casual Goodbye A handshake and a returned laptop aren’t enough to complete offboarding. Digital identities are complex, and employees accumulate access points over time, email, CRM platforms, cloud storage, social media accounts, financial software, and internal servers. Without a proper checklist, something is bound to be missed. Former accounts are prime targets for attackers. A breached personal credential might match an old work password, giving a hacker trusted access to your systems. The Information Systems Audit and Control Association (ISACA) notes that access left behind by former employees is a significant and often overlooked vulnerability. Overlooking this not only threatens your business data security but also increases compliance risk. The Pillars of a Bulletproof IT Offboarding Process A robust IT offboarding process is a strategic security measure, not just an HR task. It needs to be fast, thorough, and consistent for every departure, whether voluntary or not. The goal is to systematically remove a user’s digital footprint from your company. This process should begin before the exit interview. Close coordination between HR and IT is essential. Start with a centralised inventory of all assets and accounts the employee has. You can’t secure what you don’t know exists. Your Essential Employee Offboarding Checklist A checklist ensures nothing gets overlooked. It turns a vague intention into clear, actionable steps. Here’s a core framework you can adapt for your business: Disable network access immediately: Once an employee leaves, revoke primary login credentials, VPN access, and any remote desktop connections. Reset passwords for shared accounts: This includes social media accounts, departmental email boxes, and shared folders or workspaces. Revoke cloud access: Remove permissions for Microsoft 365, Google Workspace, Slack, project management tools, and other platforms. Using a single sign-on (SSO) portal makes it easier to manage access centrally. Reclaim all company devices: Have the employee return all company devices and perform secure data wipes before reissuing. Do not forget about mobile device management (MDM) to remotely wipe phones or tablets. Forward emails: For a smooth transition, forward the employee’s email to their manager or replacement for 30 to 90 days, then archive or delete the mailbox. You can also set an autoreply noting the departure and providing a new contact. Review and transfer digital assets: Make sure critical files aren’t stored only on personal devices, and transfer ownership of cloud documents and projects. Check access logs: Review what the employee accessed in the days before leaving. Pay attention to whether sensitive customer data was downloaded and whether it was needed for their work. The Visible Risks of Getting It Wrong The consequences of poor offboarding are very real. Data exfiltration poses serious compliance and financial risks. A departing salesperson could walk away with your entire client list, or a disgruntled developer could delete or alter critical code repositories. Even accidental data retention in personal devices and accounts could violate laws such as HIPAA and GDPR , leading to costly fines. Beyond data loss and theft, poor offboarding can also lead to financial leakage. Subscriptions to SaaS applications like Office 365, for example, may keep billing the company long after an employee has left. This is known as “ SaaS sprawl ,” and when it accumulates, it can take a real toll on your bottom line. Even if the cost is small, it’s still a sign of weak governance. Build a Culture of Secure Transitions Effective cybersecurity extends to how employees leave the company. Make the offboarding process clear from day one and include it in security training. This reinforces that access is a temporary privilege of employment, not a permanent entitlement. Documenting every step is equally important. It creates an audit trail for compliance, provides proof if issues arise, and ensures the process is repeatable and scalable as your organisation grows. Turn Employee Departures into Security Wins Treat every employee departure as a security drill and an opportunity to review access, clean up unused accounts, and reinforce your data governance policies. The goal is a thorough offboarding routine that closes gaps before they can be exploited. Don’t let former employees linger in your digital systems. A proactive, documented process is your strongest defence against this common insider threat, protecting your assets, your reputation, and your peace of mind. Contact us today to help you develop and automate a comprehensive offboarding protocol that keeps your business secure. Article FAQ What is the biggest mistake companies make during offboarding? The biggest mistake is delay. Failing to disable network and system access immediately after an employee leaves creates a window of vulnerability for data theft or misuse. Does offboarding really matter if an employee leaves on good terms? Absolutely. Even the most amicable departure poses a risk. Accounts can be hijacked, credentials can be leaked, and accidental data retention can still lead to compliance violations. Process must trump trust. What is the first IT step to take when an employee gives notice? The first step is to immediately inventory all their digital accesses and privileges alongside HR. This list drives the entire de-provisioning process and ensures nothing is missed. How can we manage offboarding for the many apps our team uses? Implement a Single Sign-On (SSO) solution. It will provide a central portal where disabling an account revokes access to all connected apps and services.
by Tanya Wetson-Catt 9 March 2026
When you first move your data and computing resources to the cloud, the bills often seem manageable. But as your business grows, a worrying trend can appear. Your cloud expenses start climbing faster than your revenue. This is not just normal growth, it is a phenomenon called cloud waste, the hidden drain on your budget hiding in your monthly cloud invoice. Cloud waste happens when you spend money on resources that do not add value to your business. Examples include underused servers, storage for completed or abandoned projects, and development or testing environments left active over the weekend. It is like keeping every piece of equipment in your factory running all the time, even when it is not needed. The cloud makes it easy to spin up resources on demand, but the same flexibility can make it easy to forget to turn them off. Most providers use a pay-as-you-go model, so the billing meter is always running. Controlling cloud waste is not just about saving money. Every dollar you save can be reinvested in innovation, stronger security, or your team. The Hidden Sources of Your Leaking Budget Cloud waste can be surprisingly easy to overlook. A common example is over-provisioning. You launch a virtual server for a project, thinking you might need a larger instance just to be safe, and then forget to scale it down. That server keeps running and billing you every hour, month after month. Orphaned resources are another common drain, especially in companies with many projects or large teams. When a project ends, do you remember to delete the storage disks, load balancers, or IP addresses that were used? Often, they stay active indefinitely. Idle resources, like databases or containers that are set up but rarely accessed, quietly add up over time. According to a 2025 report by VMWare that drew responses from over 1,800 global IT leaders, about 49% of the respondents believe that more than 25% of their public cloud expenditure is wasted, while 31% believe that waste exceeds 50%. Only 6% of the respondents believe they are not wasting any cloud spend. The FinOps Mindset: Your Financial Control Panel Fixing this level of cloud waste requires more than a one-time audit. It requires a cultural shift known as FinOps , i.e., the practice of bringing financial accountability to the variable spend model of the cloud. It is a collaborative effort where finance, technology, and business teams work together to make data-driven spending decisions. A FinOps strategy turns cloud cost from a static IT expense into a dynamic, managed business variable. The goal is not to minimise cost at all costs, but to maximise business value from every cloud dollar spent. Gaining Visibility: The Non-Negotiable First Step You can’t manage what you don’t measure, so start with the native tools your cloud provider offers. Explore their cost management consoles and take these steps to create accountability and track what’s driving expenses: Use tagging consistently to make filtering, organising, and tracking costs easier. Assign every resource to a project, department, and owner. Consider third-party cloud cost optimisation tools for deeper insights. They can automatically spot waste, recommend right-sizing actions, and consolidate data into a single dashboard if you’re using multiple cloud providers. Implementing Practical Optimisation Tactics Once you have visibility, you can act, and the easiest place to start is with the low-hanging fruit. For example: Automatically schedule non-production environments like development and testing to turn off during nights and weekends. Implement storage lifecycle policies to move old data to lower-cost archival tiers or delete it after a set period. Adjust the size of your servers by checking how much they are actually used. If the CPU is used less than 20% of the time, the server is larger than necessary, replace it with a smaller, more affordable option. Leveraging Commitments for Strategic Savings Cloud providers offer substantial discounts, like AWS Savings Plans or Azure Reserved Instances, when you commit to using a consistent level of resources for one to three years. For predictable workloads, these commitments are the most effective way to reduce unnecessary spending at full list price. The key is to make these purchases after you have right-sized your environment. Committing to an oversized instance just locks in waste. Optimise first, then commit. Making Optimisation a Continuous Cycle Managing cloud costs is not a one-time project, it’s an ongoing cycle of learning, optimising, and operating. Set up regular check-ins, monthly or quarterly, where stakeholders review cloud spending against budgets and business goals. Give your teams access to their own cost data. When developers can see the real-time impact of their architectural decisions, they become strong partners in reducing waste. Scale Smarter, Not Just Bigger The cloud offers elastic efficiency, but managing waste ensures you capture that benefit fully. It frees up capital to invest in your real business goals instead of letting it disappear into unnecessary cloud spend. As you plan for growth in 2026, make cost intelligence a core part of your strategy. Use data to guide provisioning decisions and set up automated controls to prevent waste before it starts. Reach out today for a cloud waste assessment, and we’ll help you build a sustainable FinOps practice. Article FAQ What is the most common type of cloud waste? The most common type of cloud waste is idle or underutilised compute resources, such as virtual machines, containers, or databases, that are running but not actively serving a meaningful workload, often left on accidentally or “just in case.” Can cloud waste really make a big difference to my bottom line? Absolutely. Industry reports consistently show that enterprises waste an average of 30% of their cloud spend. For a growing small business, reclaiming even 15–20% of your cloud bill can translate to thousands of dollars annually for reinvestment. Are reserved instances always the right choice to save money? They are excellent for stable, predictable workloads running 24/7. However, they are not ideal for spiky, experimental, or short-term projects. The key is to analyse your usage patterns for at least a month before making a commitment. Is automating shutdowns safe for my production systems? Automation should be applied cautiously to production. Focus initial automation efforts on non-production environments (development, testing, staging). For production, use scaling policies that automatically add/remove capacity based on real-time demand (like auto-scaling groups), which is safer than blanket shutdowns.