← Back to QA Wiki

Definition of Mock Testing

Utilizes mock objects to mimic real objects in tests.
Thank you!
Was this helpful?

Questions about Mock Testing?

Basics and Importance

  • What is mock testing?

    Mock testing involves simulating the behavior of real objects with mock objects to test the interactions between software components in isolation. Mock objects are configured to return specific values and capture calls they receive.

    In mock testing, you typically:

    1. Design the mock to mimic the behavior of the actual object.
    2. Configure expectations for the methods and properties that will be used.
    3. Execute the test by replacing the real object with the mock object.
    4. Verify that the mock object was interacted with as expected.

    To create a mock object, you might:

    const mockObject = new Mock<SomeDependency>();
    mockObject.Setup(method => method.SomeFunction()).Returns(someValue);

    Common tools include Mockito, Moq, Sinon.js, and Jest.

    For a database, you'd mock the data access layer or repository, setting up expected queries and their results:

    const mockRepository = new Mock<IDatabaseRepository>();
    mockRepository.Setup(repo => repo.Get(id)).Returns(fakeData);

    Best practices involve:

    • Clear separation of test cases.
    • Precise expectation configuration to avoid false positives/negatives.
    • Cleanup and reset of mock states between tests.

    Challenges like over-mocking or maintaining complex mock setups are mitigated by focusing on essential interactions and using factory methods for mock creation.

    To ensure effectiveness, regularly review and refactor mock tests to align with current system behavior and requirements.

    Integrate mock tests into CI/CD by including them in the test suite that runs on each build or deployment.

    In TDD, mock testing is used to test-drive the design of interfaces and interactions before implementing the actual components.

  • Why is mock testing important in software development?

    Mock testing is crucial in software development for isolating the system under test, ensuring that tests are not affected by external dependencies or stateful components. By using mock objects, developers can simulate the behavior of complex, real-world systems, which may be unavailable or difficult to configure for testing purposes. This isolation helps in pinpointing defects within the unit of code being tested, leading to more reliable and maintainable code.

    Additionally, mock testing facilitates testing in parallel, allowing different aspects of the system to be tested simultaneously without waiting for actual dependencies to be built or become available. This can significantly reduce development time and increase efficiency.

    Mocking also supports behavior verification, ensuring that the system under test interacts with its dependencies in the expected manner. This is particularly important in a service-oriented architecture where interactions between components are critical.

    Moreover, mock testing can lead to cost savings by reducing the need for setting up and maintaining complex test environments. It also allows for reproducible tests, as mock objects can be configured to return consistent results, eliminating flakiness caused by external systems.

    In summary, mock testing is a powerful technique that enhances test reliability, reduces coupling with external systems, and accelerates the development process by allowing more focused and controlled testing scenarios.

  • What are the key differences between mock testing and other types of testing?

    Mock testing differs from other types of testing in several key ways:

    • Isolation: Mocks isolate the unit of code being tested by simulating dependencies, ensuring that tests do not fail due to issues in external systems or components.

    • Control: Testers have complete control over the behavior of mock objects, allowing them to simulate various scenarios, including edge cases and error conditions that may be difficult to reproduce with real dependencies.

    • Speed: Tests using mocks run faster because they avoid the overhead of setting up and interacting with actual dependencies, such as databases or web services.

    • Determinism: Mocks provide deterministic behavior, ensuring that tests produce the same results every time they are run, which is not always the case with real dependencies that can have variable states.

    • Focus: By using mocks, tests focus solely on the code's logic rather than the integration with other systems, which is covered by integration tests.

    Here's an example of creating a mock in TypeScript using Jest:

    import { myFunction } from './myModule';
    jest.mock('./myDependency', () => {
      return {
        myDependencyFunction: jest.fn(() => 'mocked value'),
    test('myFunction calls myDependencyFunction and uses the mocked value', () => {
      expect(myFunction()).toBe('mocked value');

    In contrast, other testing types, such as integration testing, system testing, or end-to-end testing, involve working with real systems and aim to test how different parts of the application interact with each other or with the system as a whole.

  • How does mock testing improve the quality of software?

    Mock testing enhances software quality by allowing for isolated testing of components, ensuring that each part functions correctly without the interference of external dependencies. This isolation helps in identifying defects within the unit itself, rather than in the interactions with external systems, which can be tested separately in integration tests.

    By using mocks, you can simulate various scenarios, including error conditions and edge cases, that might be difficult to reproduce with actual dependencies. This thoroughness increases test coverage and improves the robustness of the software.

    Mocks also contribute to a faster and more reliable testing process. Since external systems are not involved, tests run quicker and are not prone to failures caused by issues in those systems, such as network latency or downtime.

    Furthermore, mock testing supports parallel development. Teams can work on different components simultaneously without waiting for actual implementations of the dependencies to be completed.

    Finally, mock testing can lead to better design decisions, as it often requires clear interfaces and separation of concerns to be effectively implemented. This can lead to a more modular and maintainable codebase, which is a hallmark of high-quality software.

    In summary, mock testing improves software quality by enabling isolated, thorough, and efficient testing, fostering parallel development, and encouraging good design practices.

  • What are the basic principles of mock testing?

    Mock testing relies on several basic principles to ensure effective simulation and isolation of components during testing:

    • Isolation: Mock objects are used to isolate the system under test from external dependencies or components that are not part of the current test, ensuring that tests are not affected by external factors.

    • Simulation: Mocks simulate the behavior of real objects, allowing testers to define expected interactions and outcomes, which helps in testing the system's reaction to various conditions.

    • Behavior Verification: Tests using mocks often focus on verifying that the system under test interacts with the mocks in the expected ways, such as calling methods with the correct parameters.

    • Configurability: Mocks are highly configurable, allowing testers to set up different scenarios by specifying return values, throwing exceptions, or tracking interactions.

    • Repeatability: Mock tests should be repeatable with consistent results, which is crucial for regression testing and continuous integration.

    • Simplicity: By using mocks, tests can avoid the complexity of setting up and tearing down real dependencies, leading to simpler and faster tests.

    • Focus on Unit of Work: Mock testing encourages a focus on the unit of work by testing it in isolation, which promotes better design and more maintainable code.

    Remember, mock testing should be combined with other testing methods to ensure comprehensive coverage, as it does not test the integration with real dependencies.


  • How is mock testing implemented in a software development project?

    Mock testing is implemented in a software development project through a series of strategic steps that integrate mock objects into the testing framework. Here's a concise guide:

    1. Identify dependencies in the system under test (SUT) that need to be isolated for unit testing.
    2. Design mock objects to replicate the behavior of real dependencies, adhering to the same interfaces or contracts.
    3. Configure mock objects to return expected data, simulate exceptions, or record interactions using a mocking framework like Mockito, Moq, or Sinon.js.
    // Example using Mockito in Java
    1. Inject mock objects into the SUT, often through constructor injection, setter injection, or a dependency injection framework.
    2. Write test cases that focus on the SUT's behavior, utilizing mock objects to control the test environment.
    3. Verify interactions between the SUT and mock objects to ensure correct methods are called with expected arguments.
    // Example using Mockito in Java
    1. Refactor tests as necessary when the SUT evolves, ensuring mock configurations align with new requirements.
    2. Integrate mock tests into the automated test suite to run as part of the regular build process, ensuring they contribute to CI/CD feedback loops.

    By following these steps, mock testing becomes a seamless part of the development cycle, allowing for early detection of issues and continuous validation of system behavior in isolation from external dependencies.

  • What are the steps involved in creating a mock object?

    Creating a mock object typically involves the following steps:

    1. Identify the dependency you want to replace with a mock. This could be an external service, database, or any other component your system interacts with.

    2. Define the interface or class of the dependency. Mocks are created based on the same interface that the real objects implement.

    3. Use a mocking framework to create an instance of the mock object. Popular frameworks include Mockito for Java, Moq for .NET, and Jest for JavaScript.

      MyDependency mockDependency = Mockito.mock(MyDependency.class);
    4. Configure the behavior of the mock to specify what should happen when its methods are called. This includes setting return values or throwing exceptions.

    5. Inject the mock into the system under test, replacing the real dependency. This can be done through constructor injection, setter injection, or using a dependency injection framework.

    6. Write your test to exercise the system under test, which now uses the mock object.

    7. Verify the interactions with the mock to ensure that the system under test behaves correctly. This might include checking that methods were called with the right arguments or a certain number of times.

    8. Run your test and check the results. If the test fails, investigate and correct the behavior of the system under test or update the mock configuration as needed.

  • What are some common tools used for mock testing?

    Common tools for mock testing include:

    • Mockito: A popular Java mocking framework that allows you to create and configure mock objects. Usage example:

      List mockedList = Mockito.mock(List.class);
    • Moq: Widely used in .NET for creating mock objects with a fluent API. Example:

      var mock = new Mock<IService>();
    • Sinon.js: A versatile mocking library for JavaScript, suitable for Node.js and browser environments. Example:

      const sinon = require('sinon');
      let mock = sinon.mock(myObj);
    • unittest.mock: A mocking library in Python's standard library. Example:

      from unittest.mock import MagicMock
      mock = MagicMock()
    • RSpec Mocks: A mocking framework that is part of the RSpec testing library for Ruby. Example:

      mock_model = double('Model')
    • Jest: Provides a built-in mocking library for JavaScript testing, particularly React applications. Example:

    • NSubstitute: A friendly mocking library for .NET, with a simple and clean syntax. Example:

      var substitute = Substitute.For<IService>();
    • EasyMock: Another Java mocking library that provides mock objects for interfaces. Example:

      IMockBuilder<SomeClass> builder = EasyMock.createMockBuilder(SomeClass.class);

    These tools offer various features to create, configure, and verify mock objects, helping to isolate the system under test and focus on the behavior being tested.

  • How can you create a mock test for a database?

    To create a mock test for a database, follow these steps:

    1. Identify the database operations your application performs that need to be tested.

    2. Choose a mocking framework compatible with your testing environment, such as Mockito for Java or Moq for .NET.

    3. Create a mock database interface that represents the actual database operations. This interface should mimic the behavior of the real database service.

      public interface DatabaseService {
          User getUserById(String id);
          void updateUser(User user);
    4. Implement the mock object using your chosen mocking framework. Define the expected behavior for each operation, including return values or exceptions.

      DatabaseService mockDatabase = mock(DatabaseService.class);
      when(mockDatabase.getUserById("123")).thenReturn(new User("123", "Test User"));
      doThrow(new DatabaseException()).when(mockDatabase).updateUser(any(User.class));
    5. Inject the mock object into the system under test, replacing the real database dependency.

    6. Write your test cases using the mock object to verify the system's behavior with controlled, predictable database interactions.

      public void testGetUser() {
          User user = userService.getUserById("123");
          assertEquals("Test User", user.getName());
    7. Run your tests to ensure they pass with the mock database. Adjust the mock's behavior as necessary to cover different scenarios.

    By isolating the system from the real database, you can test various data conditions and error cases without relying on an actual database, leading to faster and more reliable tests.

  • What are the best practices for implementing mock testing?

    Best practices for implementing mock testing include:

    • Designing for testability: Ensure your code is modular to easily isolate components for mocking.
    • Using clear, descriptive naming conventions: Name mocks and their methods to reflect their purpose and behavior.
    • Maintaining mocks: Keep mock implementations updated with changes in the actual components they mimic.
    • Avoiding over-mocking: Mock only what is necessary to isolate the unit of work, to prevent brittle tests.
    • Verifying interactions: Check that the system under test interacts with the mocks as expected.
    • Keeping tests focused: Each test should verify one aspect of behavior to simplify debugging when tests fail.
    • Using dependency injection: Inject mocks into the system under test to replace real dependencies.
    • Setting expectations: Define how the mock should behave before it's used, including return values and exceptions.
    • Cleaning up: Reset or dispose of mocks after each test to avoid state leakage between tests.
    • Documenting mock behavior: Comment on complex mock setups or behaviors to aid understanding for future maintainers.
    • Reviewing mock usage: Periodically review mock usage to ensure it still aligns with the actual behavior of dependencies.
    // Example of a mock setup with clear naming and behavior definition
    const mockService = {
      fetchData: jest.fn().mockResolvedValue({ data: 'mocked data' })
    // Injecting the mock service
    const systemUnderTest = new SystemUnderTest(mockService);
    // Setting expectations and verifying interactions
    • Balancing mock fidelity: Ensure mocks are accurate enough to faithfully represent the real component without unnecessary complexity.
    • Automating mock updates: Use tools that can auto-generate and update mocks based on the actual component interfaces.

Challenges and Solutions

  • What are some common challenges faced during mock testing?

    Common challenges in mock testing include:

    • Over-mocking: Excessive use of mocks can lead to tests that are brittle and overly sensitive to changes in implementation, making them difficult to maintain.
    • Complexity: Creating mocks for complex dependencies or systems with intricate behavior can be time-consuming and error-prone.
    • Behavioral Fidelity: Ensuring that mocks accurately replicate the behavior of the real objects they represent can be challenging, leading to false positives or negatives in tests.
    • Test Readability: Tests with numerous mocks or complicated setup can become hard to understand, reducing their value as documentation.
    • Integration Defects: Mocks can hide integration and interaction issues between components, which might only surface in higher-level integration tests or in production.
    • State Management: Managing the state of mocks across different test cases can be cumbersome, especially when tests are not isolated properly.
    • Tool Limitations: Mocking frameworks and tools may have limitations that prevent certain behaviors from being mocked, or they may not support the latest language features or frameworks.

    To address these challenges, apply practices such as:

    • Minimal Mocking: Only mock what is necessary to isolate the unit of work being tested.
    • Clear Abstractions: Design clear interfaces for components, making them easier to mock.
    • Incremental Testing: Complement mock testing with integration tests to catch interaction defects.
    • Test Isolation: Ensure each test is independent and manages its own mock state.
    • Documentation: Document complex mock setups to aid understanding.
    • Tool Proficiency: Stay updated with the capabilities and best practices of the chosen mocking tools.
  • How can these challenges be overcome?

    Overcoming challenges in mock testing requires a strategic approach and attention to detail. Here are some strategies:

    • Refactor Code for Testability: Ensure that the codebase is designed with testing in mind, which often means using design patterns that support dependency injection and loose coupling.

    • Use Abstraction Layers: Create abstraction layers for external services and dependencies. This allows for easier mocking and reduces the complexity of tests.

    • Invest in Quality Mocking Frameworks: Utilize robust mocking frameworks that are well-documented and widely supported. This can simplify the creation and management of mock objects.

    • Regularly Review and Update Mocks: Keep mock objects and responses up-to-date with the actual behavior of the dependencies they represent to avoid false positives or negatives.

    • Automate Mock Data Generation: Implement tools or scripts to automatically generate mock data, ensuring a diverse and realistic set of test cases.

    • Integrate Mocks into Automated Testing Pipelines: Ensure that mock tests are part of the automated testing suite and are executed as part of the CI/CD process.

    • Monitor Test Coverage: Use code coverage tools to identify areas that are not being tested and adjust mock tests accordingly.

    • Educate the Team: Provide training and resources to the team on best practices and the proper use of mock testing to avoid common pitfalls.

    • Peer Reviews: Conduct code reviews for test code, including mock tests, to catch issues early and share knowledge within the team.

    • Balance Mocking with End-to-End Tests: Complement mock tests with end-to-end tests to ensure that the system works as expected in a production-like environment.

    By implementing these strategies, test automation engineers can mitigate the challenges associated with mock testing and enhance the reliability and effectiveness of their test suites.

  • What are some common mistakes made during mock testing?

    Common mistakes during mock testing include:

    • Overusing mocks: Relying too heavily on mocks can lead to tests that are fragile and not representative of real-world scenarios. Use mocks sparingly and only when necessary.

    • Not validating interactions: Forgetting to verify that the system under test interacts with the mocks as expected can result in missed defects. Always check that the expected methods are called with the correct parameters.

    • Mocking what you don't own: Creating mocks for external dependencies not controlled by your team can lead to tests that break when those external systems change. Mock only the components you own or have control over.

    • Inadequate mock configuration: Incorrectly setting up mocks can lead to false positives or negatives. Ensure that mocks are configured to mimic the behavior of the real components accurately.

    • Ignoring side effects: Some methods have side effects that need to be replicated by the mocks. Neglecting these can lead to incomplete tests.

    • Not updating mocks: As the codebase evolves, mocks must be updated to reflect changes. Outdated mocks can cause tests to pass incorrectly or fail when they shouldn't.

    • Over-specifying mocks: Setting up mocks to expect very specific calls with exact arguments can make tests brittle. Use argument matchers to allow for some flexibility.

    • Not isolating tests: If mock setup or state is shared between tests, it can lead to inter-test dependencies and unpredictable results. Isolate each test case to ensure they run independently.

    • Lack of understanding: Misunderstanding the system being mocked can result in incorrect assumptions and ineffective tests. Gain a thorough understanding of the system before mocking it.

  • How can you ensure that your mock tests are effective?

    To ensure mock tests are effective:

    • Validate mock configurations: Ensure that mocks are set up correctly to mimic expected behavior. Incorrect configurations can lead to false positives or negatives.

    • Keep mocks up-to-date: Regularly update mocks to reflect changes in the actual dependencies they represent.

    • Verify interactions: Use verification methods to check that the system under test interacts with the mocks as expected.

    • Isolate the system under test: Ensure that the mock is the only variable in the test, to accurately assess the system's behavior.

    • Use realistic data: Mocks should return data that is representative of what the actual dependency would produce.

    • Test edge cases: Include scenarios that test how the system handles exceptional or boundary conditions through mocks.

    • Review and refactor: Periodically review mock tests to remove redundancies and improve clarity.

    • Pair with other test types: Combine mock tests with other testing methods to cover more scenarios and increase confidence in the system's reliability.

    • Automate mock tests: Integrate mock tests into your automated test suite to run them consistently and catch regressions early.

    • Monitor coverage: Use code coverage tools to ensure that mock tests are covering the intended code paths.

    • Peer review: Have mock tests reviewed by peers to catch issues that the original author may have overlooked.

    By following these guidelines, you can enhance the effectiveness of your mock tests and contribute to a more robust and reliable software testing process.

  • How can mock testing be integrated into a continuous integration/continuous deployment (CI/CD) pipeline?

    Integrating mock testing into a CI/CD pipeline involves automating the execution of mock tests as part of the build and deployment process. Here's a concise guide:

    1. Create mock tests using your preferred mocking framework.
    2. Configure your CI/CD tool to trigger the mock tests. This can be done by including a test stage in your pipeline configuration file.
    3. Use scripts to set up and tear down any required mock environments or services.
    4. Run mock tests before integration tests to ensure components behave as expected with their dependencies mocked.
    5. Analyze test results automatically. If mock tests fail, the pipeline should halt, preventing the propagation of errors.
    6. Report generation should be automated, providing visibility into the mock tests' outcomes.
    7. Maintain mock tests as part of your codebase, ensuring they evolve with your application.
      - build
      - test
      - deploy
      stage: test
        - setup-mocks.sh
        - run-mock-tests.sh
        - master

    In the above example, setup-mocks.sh would configure the necessary mock objects and services, while run-mock-tests.sh would execute the mock tests. The only directive ensures that mock tests run on the master branch, which is typically where merges occur before deployment.

Advanced Concepts

  • What is the role of mock testing in Test Driven Development (TDD)?

    In Test Driven Development (TDD), mock testing plays a pivotal role by simulating the behavior of real objects that are yet to be implemented or are impractical to include in unit tests. By using mocks, developers can focus on the unit under test, ensuring that tests are isolated and not dependent on external systems or complex dependencies.

    During the TDD cycle of writing a failing test, implementing the minimal code to pass the test, and then refactoring, mocks are often introduced in the first phase. They help in specifying the expected interactions with dependencies, which drives the design of the interfaces. This is particularly useful when the actual implementation of these dependencies might be time-consuming or would introduce flakiness in the tests.

    Mocks enable developers to verify that the correct methods are called with the expected parameters, which is crucial for the contract between different parts of the system. This allows for a design that is more modular and adheres to the Single Responsibility Principle.

    Furthermore, by using mocks, the feedback loop in TDD is significantly shortened, as there is no need to wait for actual dependencies to respond or be available. This accelerates the development process and helps maintain a steady pace.

    // Example of using a mock in TDD (TypeScript)
    test('should call dependency method with correct parameters', () => {
      const mockDependency = {
        performAction: jest.fn()
      const systemUnderTest = new SystemUnderTest(mockDependency);

    In summary, mock testing within TDD ensures that each unit can be tested in isolation, supports better design, and speeds up the development cycle.

  • How does mock testing work in a microservices architecture?

    In a microservices architecture, mock testing involves simulating the behavior of external services that a microservice interacts with. This allows for isolated testing of the service in question.

    To implement mock testing in microservices:

    1. Identify the external dependencies of the microservice.
    2. Create mock objects or stubs for these dependencies using a mocking framework or tool.
    3. Configure the microservice to use these mocks instead of the actual services during the test execution.
    4. Write test cases that exercise the functionality of the microservice, asserting that it behaves correctly with the mocked interfaces.

    Mocking in microservices is particularly useful for:

    • Testing error handling by simulating failures of dependencies.
    • Parallel development where dependent services are not yet available or are being developed concurrently.
    • Continuous Integration to ensure that tests can run independently of the environment.

    Example using a JavaScript mocking library:

    const { myMicroservice } = require('./myMicroservice');
    const { mockDependencyService } = require('mocking-library');
    test('should handle dependency failure gracefully', () => {
        endpoint: '/external-service',
        method: 'GET',
        response: { status: 500 }
      const response = myMicroservice.handleExternalService();
      expect(response).toEqual('Error handling logic executed');

    Best practices include:

    • Ensuring mocks accurately reflect the behavior of real dependencies.
    • Keeping mock configurations up-to-date with service contracts.
    • Using contract testing to validate that mocks and actual services agree on the expected interactions.
  • What is the concept of 'stubbing' in mock testing?

    Stubbing is a technique used in mock testing to replace parts of the system under test with simplified implementations that return predetermined responses. It's a form of test double that stands in for real functionality that is either not implemented or would be impractical to use during testing due to side effects, slowness, or non-determinism.

    Unlike full-fledged mocks, stubs typically do not have expectations about how they are called. They are primarily used to control the behavior of the system under test by returning specific values or throwing exceptions, thus allowing tests to focus on the code paths that are triggered by those responses.

    Here's an example in TypeScript using a popular stubbing library, Sinon.js:

    import sinon from 'sinon';
    import { MyService } from './my-service';
    import { expect } from 'chai';
    describe('MyService', () => {
      it('should handle the response from a stubbed dependency', () => {
        const myService = new MyService();
        const stub = sinon.stub(myService, 'dependencyMethod').returns('stubbed value');
        const result = myService.callDependencyMethod();
        expect(result).to.equal('stubbed value');

    In this example, dependencyMethod is stubbed to return 'stubbed value' whenever it is called within MyService. This allows the test to verify behavior without relying on the actual implementation of dependencyMethod.

    Stubbing is particularly useful when dealing with external services, database calls, or any other components that would introduce complexity or non-determinism into the tests. It helps create isolated and predictable test environments where the focus is on the unit of code being tested.

  • How can mock testing be used for performance testing?

    Mock testing can be leveraged in performance testing to simulate the behavior of components that are either unavailable or too costly to include in a full-scale performance test. By replacing these components with mocks, you can isolate and test the performance of specific parts of the system under load without the overhead or interference from external systems.

    For instance, if you're testing an application that relies on a third-party service, you can use a mock to mimic the latency and throughput of the real service. This allows you to:

    • Control the test environment more predictably by eliminating external dependencies that could introduce variability.
    • Simulate various scenarios, such as high latency or low bandwidth, to understand how your application behaves under different conditions.
    • Stress test individual components by simulating high loads that might not be possible or practical with the actual dependent service.

    Here's an example of creating a mock for a service in a performance test scenario:

    // Mock service that simulates a delay
    function mockService(delay) {
      return new Promise((resolve) => setTimeout(resolve, delay));
    // Performance test using the mock service
    async function performanceTest() {
      const startTime = performance.now();
      await mockService(100); // Simulates a 100ms delay
      const endTime = performance.now();
      console.log(`Service call took ${endTime - startTime} milliseconds`);

    In this code, mockService simulates a service call with a specified delay, and performanceTest measures the time taken to complete the call. By adjusting the delay, you can test how your system handles different response times.

    Using mocks in performance testing is a cost-effective and flexible approach to identify bottlenecks and optimize the performance of the system under test.

  • What is the difference between a mock and a spy in testing?

    In test automation, mocks and spies serve different purposes when isolating units of code for testing. A mock is an object that replaces a real component, with pre-programmed behaviors and expectations set. It's used to verify interactions between the system under test and the mock.

    // Example of a mock
    const mockObject = {
      method: jest.fn().mockReturnValue("mocked value")

    A spy, on the other hand, is used to wrap an existing function, allowing the test to record information about how the function is used without altering its behavior. Spies can track function calls, arguments, and return values, and they can also change the behavior of the function if needed.

    // Example of a spy
    const spy = jest.spyOn(realObject, 'method');

    The key difference lies in their usage:

    • Mocks are about creating a fake version of an interface or class with preset behavior, and they're typically used when the actual implementation is irrelevant or when it needs to be controlled for the test scenario.
    • Spies are about gathering information on how a function is used during the execution of the test. They are useful when you need to assert that certain functions are called with the right arguments or the correct number of times, without changing the actual implementation.

    Both are valuable in different contexts, with mocks being more about controlling the external dependencies and spies being more about observing them.

AboutQuestionsDiscord ForumBrowser ExtensionTagsQA Jobs

Rayrun is a community for QA engineers. I am constantly looking for new ways to add value to people learning Playwright and other browser automation frameworks. If you have feedback, email luc@ray.run.