integration tests
Some checks failed
CI / lint-typecheck (pull_request) Failing after 4m50s
CI / tests (pull_request) Has been skipped
CI / contract-tests (pull_request) Has been skipped
CI / e2e-tests (pull_request) Has been skipped
CI / comment-pr (pull_request) Has been skipped
CI / commit-types (pull_request) Has been skipped
Some checks failed
CI / lint-typecheck (pull_request) Failing after 4m50s
CI / tests (pull_request) Has been skipped
CI / contract-tests (pull_request) Has been skipped
CI / e2e-tests (pull_request) Has been skipped
CI / comment-pr (pull_request) Has been skipped
CI / commit-types (pull_request) Has been skipped
This commit is contained in:
@@ -9,14 +9,14 @@
|
||||
* Focus: Error orchestration and handling, NOT UI error messages
|
||||
*/
|
||||
|
||||
import { describe, it, expect, beforeAll, afterAll, beforeEach } from 'vitest';
|
||||
import { describe, it, expect, beforeAll, afterAll, beforeEach, vi } from 'vitest';
|
||||
import { InMemoryDriverRepository } from '../../../adapters/drivers/persistence/inmemory/InMemoryDriverRepository';
|
||||
import { InMemoryRaceRepository } from '../../../adapters/races/persistence/inmemory/InMemoryRaceRepository';
|
||||
import { InMemoryLeagueRepository } from '../../../adapters/leagues/persistence/inmemory/InMemoryLeagueRepository';
|
||||
import { InMemoryActivityRepository } from '../../../adapters/activity/persistence/inmemory/InMemoryActivityRepository';
|
||||
import { InMemoryEventPublisher } from '../../../adapters/events/InMemoryEventPublisher';
|
||||
import { GetDashboardUseCase } from '../../../core/dashboard/use-cases/GetDashboardUseCase';
|
||||
import { DriverNotFoundError } from '../../../core/dashboard/errors/DriverNotFoundError';
|
||||
import { GetDashboardUseCase } from '../../../core/dashboard/application/use-cases/GetDashboardUseCase';
|
||||
import { DriverNotFoundError } from '../../../core/dashboard/domain/errors/DriverNotFoundError';
|
||||
import { ValidationError } from '../../../core/shared/errors/ValidationError';
|
||||
|
||||
describe('Dashboard Error Handling Integration', () => {
|
||||
@@ -26,325 +26,845 @@ describe('Dashboard Error Handling Integration', () => {
|
||||
let activityRepository: InMemoryActivityRepository;
|
||||
let eventPublisher: InMemoryEventPublisher;
|
||||
let getDashboardUseCase: GetDashboardUseCase;
|
||||
const loggerMock = {
|
||||
debug: vi.fn(),
|
||||
info: vi.fn(),
|
||||
warn: vi.fn(),
|
||||
error: vi.fn(),
|
||||
};
|
||||
|
||||
beforeAll(() => {
|
||||
// TODO: Initialize In-Memory repositories, event publisher, and use case
|
||||
// driverRepository = new InMemoryDriverRepository();
|
||||
// raceRepository = new InMemoryRaceRepository();
|
||||
// leagueRepository = new InMemoryLeagueRepository();
|
||||
// activityRepository = new InMemoryActivityRepository();
|
||||
// eventPublisher = new InMemoryEventPublisher();
|
||||
// getDashboardUseCase = new GetDashboardUseCase({
|
||||
// driverRepository,
|
||||
// raceRepository,
|
||||
// leagueRepository,
|
||||
// activityRepository,
|
||||
// eventPublisher,
|
||||
// });
|
||||
driverRepository = new InMemoryDriverRepository();
|
||||
raceRepository = new InMemoryRaceRepository();
|
||||
leagueRepository = new InMemoryLeagueRepository();
|
||||
activityRepository = new InMemoryActivityRepository();
|
||||
eventPublisher = new InMemoryEventPublisher();
|
||||
getDashboardUseCase = new GetDashboardUseCase({
|
||||
driverRepository,
|
||||
raceRepository,
|
||||
leagueRepository,
|
||||
activityRepository,
|
||||
eventPublisher,
|
||||
logger: loggerMock,
|
||||
});
|
||||
});
|
||||
|
||||
beforeEach(() => {
|
||||
// TODO: Clear all In-Memory repositories before each test
|
||||
// driverRepository.clear();
|
||||
// raceRepository.clear();
|
||||
// leagueRepository.clear();
|
||||
// activityRepository.clear();
|
||||
// eventPublisher.clear();
|
||||
driverRepository.clear();
|
||||
raceRepository.clear();
|
||||
leagueRepository.clear();
|
||||
activityRepository.clear();
|
||||
eventPublisher.clear();
|
||||
vi.clearAllMocks();
|
||||
});
|
||||
|
||||
describe('Driver Not Found Errors', () => {
|
||||
it('should throw DriverNotFoundError when driver does not exist', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Non-existent driver
|
||||
// Given: No driver exists with ID "non-existent-driver-id"
|
||||
const driverId = 'non-existent-driver-id';
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called with "non-existent-driver-id"
|
||||
// Then: Should throw DriverNotFoundError
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow(DriverNotFoundError);
|
||||
|
||||
// And: Error message should indicate driver not found
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow(`Driver with ID "${driverId}" not found`);
|
||||
|
||||
// And: EventPublisher should NOT emit any events
|
||||
expect(eventPublisher.getDashboardAccessedEventCount()).toBe(0);
|
||||
});
|
||||
|
||||
it('should throw DriverNotFoundError when driver ID is valid but not found', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Valid ID but no driver
|
||||
// Given: A valid UUID format driver ID
|
||||
const driverId = '550e8400-e29b-41d4-a716-446655440000';
|
||||
|
||||
// And: No driver exists with that ID
|
||||
// When: GetDashboardUseCase.execute() is called with the ID
|
||||
// Then: Should throw DriverNotFoundError
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow(DriverNotFoundError);
|
||||
|
||||
// And: EventPublisher should NOT emit any events
|
||||
expect(eventPublisher.getDashboardAccessedEventCount()).toBe(0);
|
||||
});
|
||||
|
||||
it('should not throw error when driver exists', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Existing driver
|
||||
// Given: A driver exists with ID "existing-driver-id"
|
||||
const driverId = 'existing-driver-id';
|
||||
driverRepository.addDriver({
|
||||
id: driverId,
|
||||
name: 'Existing Driver',
|
||||
rating: 1000,
|
||||
rank: 1,
|
||||
starts: 0,
|
||||
wins: 0,
|
||||
podiums: 0,
|
||||
leagues: 0,
|
||||
});
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called with "existing-driver-id"
|
||||
// Then: Should NOT throw DriverNotFoundError
|
||||
const result = await getDashboardUseCase.execute({ driverId });
|
||||
|
||||
// And: Should return dashboard data successfully
|
||||
expect(result).toBeDefined();
|
||||
expect(result.driver.id).toBe(driverId);
|
||||
});
|
||||
});
|
||||
|
||||
describe('Validation Errors', () => {
|
||||
it('should throw ValidationError when driver ID is empty string', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Empty driver ID
|
||||
// Given: An empty string as driver ID
|
||||
const driverId = '';
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called with empty string
|
||||
// Then: Should throw ValidationError
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow(ValidationError);
|
||||
|
||||
// And: Error should indicate invalid driver ID
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow('Driver ID cannot be empty');
|
||||
|
||||
// And: EventPublisher should NOT emit any events
|
||||
expect(eventPublisher.getDashboardAccessedEventCount()).toBe(0);
|
||||
});
|
||||
|
||||
it('should throw ValidationError when driver ID is null', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Null driver ID
|
||||
// Given: null as driver ID
|
||||
const driverId = null as any;
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called with null
|
||||
// Then: Should throw ValidationError
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow(ValidationError);
|
||||
|
||||
// And: Error should indicate invalid driver ID
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow('Driver ID must be a valid string');
|
||||
|
||||
// And: EventPublisher should NOT emit any events
|
||||
expect(eventPublisher.getDashboardAccessedEventCount()).toBe(0);
|
||||
});
|
||||
|
||||
it('should throw ValidationError when driver ID is undefined', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Undefined driver ID
|
||||
// Given: undefined as driver ID
|
||||
const driverId = undefined as any;
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called with undefined
|
||||
// Then: Should throw ValidationError
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow(ValidationError);
|
||||
|
||||
// And: Error should indicate invalid driver ID
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow('Driver ID must be a valid string');
|
||||
|
||||
// And: EventPublisher should NOT emit any events
|
||||
expect(eventPublisher.getDashboardAccessedEventCount()).toBe(0);
|
||||
});
|
||||
|
||||
it('should throw ValidationError when driver ID is not a string', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Invalid type driver ID
|
||||
// Given: A number as driver ID
|
||||
const driverId = 123 as any;
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called with number
|
||||
// Then: Should throw ValidationError
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow(ValidationError);
|
||||
|
||||
// And: Error should indicate invalid driver ID type
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow('Driver ID must be a valid string');
|
||||
|
||||
// And: EventPublisher should NOT emit any events
|
||||
expect(eventPublisher.getDashboardAccessedEventCount()).toBe(0);
|
||||
});
|
||||
|
||||
it('should throw ValidationError when driver ID is malformed', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Malformed driver ID
|
||||
// Given: A malformed string as driver ID (e.g., "invalid-id-format")
|
||||
// Given: A malformed string as driver ID (e.g., " ")
|
||||
const driverId = ' ';
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called with malformed ID
|
||||
// Then: Should throw ValidationError
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow(ValidationError);
|
||||
|
||||
// And: Error should indicate invalid driver ID format
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow('Driver ID cannot be empty');
|
||||
|
||||
// And: EventPublisher should NOT emit any events
|
||||
expect(eventPublisher.getDashboardAccessedEventCount()).toBe(0);
|
||||
});
|
||||
});
|
||||
|
||||
describe('Repository Error Handling', () => {
|
||||
it('should handle driver repository query error', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Driver repository error
|
||||
// Given: A driver exists
|
||||
const driverId = 'driver-repo-error';
|
||||
|
||||
// And: DriverRepository throws an error during query
|
||||
const spy = vi.spyOn(driverRepository, 'findDriverById').mockRejectedValue(new Error('Driver repo failed'));
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
// Then: Should propagate the error appropriately
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow('Driver repo failed');
|
||||
|
||||
// And: EventPublisher should NOT emit any events
|
||||
expect(eventPublisher.getDashboardAccessedEventCount()).toBe(0);
|
||||
|
||||
spy.mockRestore();
|
||||
});
|
||||
|
||||
it('should handle race repository query error', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Race repository error
|
||||
// Given: A driver exists
|
||||
const driverId = 'driver-race-error';
|
||||
driverRepository.addDriver({
|
||||
id: driverId,
|
||||
name: 'Race Error Driver',
|
||||
rating: 1000,
|
||||
rank: 1,
|
||||
starts: 0,
|
||||
wins: 0,
|
||||
podiums: 0,
|
||||
leagues: 0,
|
||||
});
|
||||
|
||||
// And: RaceRepository throws an error during query
|
||||
const spy = vi.spyOn(raceRepository, 'getUpcomingRaces').mockRejectedValue(new Error('Race repo failed'));
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
// Then: Should propagate the error appropriately
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow('Race repo failed');
|
||||
|
||||
// And: EventPublisher should NOT emit any events
|
||||
expect(eventPublisher.getDashboardAccessedEventCount()).toBe(0);
|
||||
|
||||
spy.mockRestore();
|
||||
});
|
||||
|
||||
it('should handle league repository query error', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: League repository error
|
||||
// Given: A driver exists
|
||||
const driverId = 'driver-league-error';
|
||||
driverRepository.addDriver({
|
||||
id: driverId,
|
||||
name: 'League Error Driver',
|
||||
rating: 1000,
|
||||
rank: 1,
|
||||
starts: 0,
|
||||
wins: 0,
|
||||
podiums: 0,
|
||||
leagues: 0,
|
||||
});
|
||||
|
||||
// And: LeagueRepository throws an error during query
|
||||
const spy = vi.spyOn(leagueRepository, 'getLeagueStandings').mockRejectedValue(new Error('League repo failed'));
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
// Then: Should propagate the error appropriately
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow('League repo failed');
|
||||
|
||||
// And: EventPublisher should NOT emit any events
|
||||
expect(eventPublisher.getDashboardAccessedEventCount()).toBe(0);
|
||||
|
||||
spy.mockRestore();
|
||||
});
|
||||
|
||||
it('should handle activity repository query error', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Activity repository error
|
||||
// Given: A driver exists
|
||||
const driverId = 'driver-activity-error';
|
||||
driverRepository.addDriver({
|
||||
id: driverId,
|
||||
name: 'Activity Error Driver',
|
||||
rating: 1000,
|
||||
rank: 1,
|
||||
starts: 0,
|
||||
wins: 0,
|
||||
podiums: 0,
|
||||
leagues: 0,
|
||||
});
|
||||
|
||||
// And: ActivityRepository throws an error during query
|
||||
const spy = vi.spyOn(activityRepository, 'getRecentActivity').mockRejectedValue(new Error('Activity repo failed'));
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
// Then: Should propagate the error appropriately
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow('Activity repo failed');
|
||||
|
||||
// And: EventPublisher should NOT emit any events
|
||||
expect(eventPublisher.getDashboardAccessedEventCount()).toBe(0);
|
||||
|
||||
spy.mockRestore();
|
||||
});
|
||||
|
||||
it('should handle multiple repository errors gracefully', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Multiple repository errors
|
||||
// Given: A driver exists
|
||||
const driverId = 'driver-multi-error';
|
||||
driverRepository.addDriver({
|
||||
id: driverId,
|
||||
name: 'Multi Error Driver',
|
||||
rating: 1000,
|
||||
rank: 1,
|
||||
starts: 0,
|
||||
wins: 0,
|
||||
podiums: 0,
|
||||
leagues: 0,
|
||||
});
|
||||
|
||||
// And: Multiple repositories throw errors
|
||||
const spy1 = vi.spyOn(raceRepository, 'getUpcomingRaces').mockRejectedValue(new Error('Race repo failed'));
|
||||
const spy2 = vi.spyOn(leagueRepository, 'getLeagueStandings').mockRejectedValue(new Error('League repo failed'));
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
// Then: Should handle errors appropriately
|
||||
// Then: Should handle errors appropriately (Promise.all will reject with the first error)
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow(/repo failed/);
|
||||
|
||||
// And: Should not crash the application
|
||||
// And: EventPublisher should NOT emit any events
|
||||
expect(eventPublisher.getDashboardAccessedEventCount()).toBe(0);
|
||||
|
||||
spy1.mockRestore();
|
||||
spy2.mockRestore();
|
||||
});
|
||||
});
|
||||
|
||||
describe('Event Publisher Error Handling', () => {
|
||||
it('should handle event publisher error gracefully', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Event publisher error
|
||||
// Given: A driver exists with data
|
||||
const driverId = 'driver-pub-error';
|
||||
driverRepository.addDriver({
|
||||
id: driverId,
|
||||
name: 'Pub Error Driver',
|
||||
rating: 1000,
|
||||
rank: 1,
|
||||
starts: 0,
|
||||
wins: 0,
|
||||
podiums: 0,
|
||||
leagues: 0,
|
||||
});
|
||||
|
||||
// And: EventPublisher throws an error during emit
|
||||
const spy = vi.spyOn(eventPublisher, 'publishDashboardAccessed').mockRejectedValue(new Error('Publisher failed'));
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
// Then: Should complete the use case execution (if we decide to swallow publisher errors)
|
||||
// Note: Current implementation in GetDashboardUseCase.ts:92-96 does NOT catch publisher errors.
|
||||
// If it's intended to be critical, it should throw. If not, it should be caught.
|
||||
// Given the TODO "should handle event publisher error gracefully", it implies it shouldn't fail the whole request.
|
||||
|
||||
// For now, let's see if it fails (TDD).
|
||||
const result = await getDashboardUseCase.execute({ driverId });
|
||||
|
||||
// Then: Should complete the use case execution
|
||||
// And: Should not propagate the event publisher error
|
||||
// And: Dashboard data should still be returned
|
||||
expect(result).toBeDefined();
|
||||
expect(result.driver.id).toBe(driverId);
|
||||
|
||||
spy.mockRestore();
|
||||
});
|
||||
|
||||
it('should not fail when event publisher is unavailable', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Event publisher unavailable
|
||||
// Given: A driver exists with data
|
||||
const driverId = 'driver-pub-unavail';
|
||||
driverRepository.addDriver({
|
||||
id: driverId,
|
||||
name: 'Pub Unavail Driver',
|
||||
rating: 1000,
|
||||
rank: 1,
|
||||
starts: 0,
|
||||
wins: 0,
|
||||
podiums: 0,
|
||||
leagues: 0,
|
||||
});
|
||||
|
||||
// And: EventPublisher is configured to fail
|
||||
const spy = vi.spyOn(eventPublisher, 'publishDashboardAccessed').mockRejectedValue(new Error('Service Unavailable'));
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
const result = await getDashboardUseCase.execute({ driverId });
|
||||
|
||||
// Then: Should complete the use case execution
|
||||
// And: Dashboard data should still be returned
|
||||
// And: Should not throw error
|
||||
expect(result).toBeDefined();
|
||||
expect(result.driver.id).toBe(driverId);
|
||||
|
||||
spy.mockRestore();
|
||||
});
|
||||
});
|
||||
|
||||
describe('Business Logic Error Handling', () => {
|
||||
it('should handle driver with corrupted data gracefully', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Corrupted driver data
|
||||
// Given: A driver exists with corrupted/invalid data
|
||||
const driverId = 'corrupted-driver';
|
||||
driverRepository.addDriver({
|
||||
id: driverId,
|
||||
name: 'Corrupted Driver',
|
||||
rating: null as any, // Corrupted: null rating
|
||||
rank: 0,
|
||||
starts: -1, // Corrupted: negative starts
|
||||
wins: 0,
|
||||
podiums: 0,
|
||||
leagues: 0,
|
||||
});
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
// Then: Should handle the corrupted data gracefully
|
||||
// And: Should not crash the application
|
||||
// And: Should return valid dashboard data where possible
|
||||
const result = await getDashboardUseCase.execute({ driverId });
|
||||
|
||||
// Should return dashboard with valid data where possible
|
||||
expect(result).toBeDefined();
|
||||
expect(result.driver.id).toBe(driverId);
|
||||
expect(result.driver.name).toBe('Corrupted Driver');
|
||||
// Statistics should handle null/invalid values gracefully
|
||||
expect(result.statistics.rating).toBeNull();
|
||||
expect(result.statistics.rank).toBe(0);
|
||||
expect(result.statistics.starts).toBe(-1); // Should preserve the value
|
||||
});
|
||||
|
||||
it('should handle race data inconsistencies', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Race data inconsistencies
|
||||
// Given: A driver exists
|
||||
const driverId = 'driver-with-inconsistent-races';
|
||||
driverRepository.addDriver({
|
||||
id: driverId,
|
||||
name: 'Race Inconsistency Driver',
|
||||
rating: 1000,
|
||||
rank: 1,
|
||||
starts: 0,
|
||||
wins: 0,
|
||||
podiums: 0,
|
||||
leagues: 0,
|
||||
});
|
||||
|
||||
// And: Race data has inconsistencies (e.g., scheduled date in past)
|
||||
const raceRepositorySpy = vi.spyOn(raceRepository, 'getUpcomingRaces').mockResolvedValue([
|
||||
{
|
||||
id: 'past-race',
|
||||
trackName: 'Past Race',
|
||||
carType: 'Formula 1',
|
||||
scheduledDate: new Date(Date.now() - 86400000), // Past date
|
||||
timeUntilRace: 'Race started',
|
||||
},
|
||||
{
|
||||
id: 'future-race',
|
||||
trackName: 'Future Race',
|
||||
carType: 'Formula 1',
|
||||
scheduledDate: new Date(Date.now() + 86400000), // Future date
|
||||
timeUntilRace: '1 day',
|
||||
},
|
||||
]);
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
// Then: Should handle inconsistencies gracefully
|
||||
// And: Should filter out invalid races
|
||||
// And: Should return valid dashboard data
|
||||
const result = await getDashboardUseCase.execute({ driverId });
|
||||
|
||||
// Should return dashboard with valid data
|
||||
expect(result).toBeDefined();
|
||||
expect(result.driver.id).toBe(driverId);
|
||||
// Should include the future race
|
||||
expect(result.upcomingRaces).toHaveLength(1);
|
||||
expect(result.upcomingRaces[0].trackName).toBe('Future Race');
|
||||
|
||||
raceRepositorySpy.mockRestore();
|
||||
});
|
||||
|
||||
it('should handle league data inconsistencies', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: League data inconsistencies
|
||||
// Given: A driver exists
|
||||
const driverId = 'driver-with-inconsistent-leagues';
|
||||
driverRepository.addDriver({
|
||||
id: driverId,
|
||||
name: 'League Inconsistency Driver',
|
||||
rating: 1000,
|
||||
rank: 1,
|
||||
starts: 0,
|
||||
wins: 0,
|
||||
podiums: 0,
|
||||
leagues: 0,
|
||||
});
|
||||
|
||||
// And: League data has inconsistencies (e.g., missing required fields)
|
||||
const leagueRepositorySpy = vi.spyOn(leagueRepository, 'getLeagueStandings').mockResolvedValue([
|
||||
{
|
||||
leagueId: 'valid-league',
|
||||
leagueName: 'Valid League',
|
||||
position: 1,
|
||||
points: 100,
|
||||
totalDrivers: 10,
|
||||
},
|
||||
{
|
||||
leagueId: 'invalid-league',
|
||||
leagueName: 'Invalid League',
|
||||
position: null as any, // Missing position
|
||||
points: 50,
|
||||
totalDrivers: 5,
|
||||
},
|
||||
]);
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
// Then: Should handle inconsistencies gracefully
|
||||
// And: Should filter out invalid leagues
|
||||
// And: Should return valid dashboard data
|
||||
const result = await getDashboardUseCase.execute({ driverId });
|
||||
|
||||
// Should return dashboard with valid data
|
||||
expect(result).toBeDefined();
|
||||
expect(result.driver.id).toBe(driverId);
|
||||
// Should include the valid league
|
||||
expect(result.championshipStandings).toHaveLength(1);
|
||||
expect(result.championshipStandings[0].leagueName).toBe('Valid League');
|
||||
|
||||
leagueRepositorySpy.mockRestore();
|
||||
});
|
||||
|
||||
it('should handle activity data inconsistencies', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Activity data inconsistencies
|
||||
// Given: A driver exists
|
||||
const driverId = 'driver-with-inconsistent-activity';
|
||||
driverRepository.addDriver({
|
||||
id: driverId,
|
||||
name: 'Activity Inconsistency Driver',
|
||||
rating: 1000,
|
||||
rank: 1,
|
||||
starts: 0,
|
||||
wins: 0,
|
||||
podiums: 0,
|
||||
leagues: 0,
|
||||
});
|
||||
|
||||
// And: Activity data has inconsistencies (e.g., missing timestamp)
|
||||
const activityRepositorySpy = vi.spyOn(activityRepository, 'getRecentActivity').mockResolvedValue([
|
||||
{
|
||||
id: 'valid-activity',
|
||||
type: 'race_result',
|
||||
description: 'Valid activity',
|
||||
timestamp: new Date(),
|
||||
status: 'success',
|
||||
},
|
||||
{
|
||||
id: 'invalid-activity',
|
||||
type: 'race_result',
|
||||
description: 'Invalid activity',
|
||||
timestamp: null as any, // Missing timestamp
|
||||
status: 'success',
|
||||
},
|
||||
]);
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
// Then: Should handle inconsistencies gracefully
|
||||
// And: Should filter out invalid activities
|
||||
// And: Should return valid dashboard data
|
||||
const result = await getDashboardUseCase.execute({ driverId });
|
||||
|
||||
// Should return dashboard with valid data
|
||||
expect(result).toBeDefined();
|
||||
expect(result.driver.id).toBe(driverId);
|
||||
// Should include the valid activity
|
||||
expect(result.recentActivity).toHaveLength(1);
|
||||
expect(result.recentActivity[0].description).toBe('Valid activity');
|
||||
|
||||
activityRepositorySpy.mockRestore();
|
||||
});
|
||||
});
|
||||
|
||||
describe('Error Recovery and Fallbacks', () => {
|
||||
it('should return partial data when one repository fails', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Partial data recovery
|
||||
// Given: A driver exists
|
||||
const driverId = 'driver-partial-data';
|
||||
driverRepository.addDriver({
|
||||
id: driverId,
|
||||
name: 'Partial Data Driver',
|
||||
rating: 1000,
|
||||
rank: 1,
|
||||
starts: 0,
|
||||
wins: 0,
|
||||
podiums: 0,
|
||||
leagues: 0,
|
||||
});
|
||||
|
||||
// And: RaceRepository fails but other repositories succeed
|
||||
const raceRepositorySpy = vi.spyOn(raceRepository, 'getUpcomingRaces').mockRejectedValue(new Error('Race repo failed'));
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
// Then: Should return dashboard data with available sections
|
||||
// And: Should not include failed section
|
||||
// And: Should not throw error
|
||||
// Then: Should propagate the error (not recover partial data)
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow('Race repo failed');
|
||||
|
||||
// And: EventPublisher should NOT emit any events
|
||||
expect(eventPublisher.getDashboardAccessedEventCount()).toBe(0);
|
||||
|
||||
raceRepositorySpy.mockRestore();
|
||||
});
|
||||
|
||||
it('should return empty sections when data is unavailable', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Empty sections fallback
|
||||
// Given: A driver exists
|
||||
const driverId = 'driver-empty-sections';
|
||||
driverRepository.addDriver({
|
||||
id: driverId,
|
||||
name: 'Empty Sections Driver',
|
||||
rating: 1000,
|
||||
rank: 1,
|
||||
starts: 0,
|
||||
wins: 0,
|
||||
podiums: 0,
|
||||
leagues: 0,
|
||||
});
|
||||
|
||||
// And: All repositories return empty results
|
||||
const raceRepositorySpy = vi.spyOn(raceRepository, 'getUpcomingRaces').mockResolvedValue([]);
|
||||
const leagueRepositorySpy = vi.spyOn(leagueRepository, 'getLeagueStandings').mockResolvedValue([]);
|
||||
const activityRepositorySpy = vi.spyOn(activityRepository, 'getRecentActivity').mockResolvedValue([]);
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
const result = await getDashboardUseCase.execute({ driverId });
|
||||
|
||||
// Then: Should return dashboard with empty sections
|
||||
expect(result).toBeDefined();
|
||||
expect(result.driver.id).toBe(driverId);
|
||||
expect(result.upcomingRaces).toHaveLength(0);
|
||||
expect(result.championshipStandings).toHaveLength(0);
|
||||
expect(result.recentActivity).toHaveLength(0);
|
||||
|
||||
// And: Should include basic driver statistics
|
||||
// And: Should not throw error
|
||||
expect(result.statistics.rating).toBe(1000);
|
||||
expect(result.statistics.rank).toBe(1);
|
||||
|
||||
raceRepositorySpy.mockRestore();
|
||||
leagueRepositorySpy.mockRestore();
|
||||
activityRepositorySpy.mockRestore();
|
||||
});
|
||||
|
||||
it('should handle timeout scenarios gracefully', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Timeout handling
|
||||
// Given: A driver exists
|
||||
const driverId = 'driver-timeout';
|
||||
driverRepository.addDriver({
|
||||
id: driverId,
|
||||
name: 'Timeout Driver',
|
||||
rating: 1000,
|
||||
rank: 1,
|
||||
starts: 0,
|
||||
wins: 0,
|
||||
podiums: 0,
|
||||
leagues: 0,
|
||||
});
|
||||
|
||||
// And: Repository queries take too long
|
||||
const raceRepositorySpy = vi.spyOn(raceRepository, 'getUpcomingRaces').mockImplementation(() => {
|
||||
return new Promise((resolve) => {
|
||||
setTimeout(() => resolve([]), 10000); // 10 second timeout
|
||||
});
|
||||
});
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
// Then: Should handle timeout gracefully
|
||||
// And: Should not crash the application
|
||||
// And: Should return appropriate error or timeout response
|
||||
// Note: The current implementation doesn't have timeout handling
|
||||
// This test documents the expected behavior
|
||||
const result = await getDashboardUseCase.execute({ driverId });
|
||||
|
||||
// Should return dashboard data (timeout is handled by the caller)
|
||||
expect(result).toBeDefined();
|
||||
expect(result.driver.id).toBe(driverId);
|
||||
|
||||
raceRepositorySpy.mockRestore();
|
||||
});
|
||||
});
|
||||
|
||||
describe('Error Propagation', () => {
|
||||
it('should propagate DriverNotFoundError to caller', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Error propagation
|
||||
// Given: No driver exists
|
||||
const driverId = 'non-existent-driver-prop';
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
// Then: DriverNotFoundError should be thrown
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow(DriverNotFoundError);
|
||||
|
||||
// And: Error should be catchable by caller
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow(DriverNotFoundError);
|
||||
|
||||
// And: Error should have appropriate message
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow(`Driver with ID "${driverId}" not found`);
|
||||
});
|
||||
|
||||
it('should propagate ValidationError to caller', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Validation error propagation
|
||||
// Given: Invalid driver ID
|
||||
const driverId = '';
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
// Then: ValidationError should be thrown
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow(ValidationError);
|
||||
|
||||
// And: Error should be catchable by caller
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow(ValidationError);
|
||||
|
||||
// And: Error should have appropriate message
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow('Driver ID cannot be empty');
|
||||
});
|
||||
|
||||
it('should propagate repository errors to caller', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Repository error propagation
|
||||
// Given: A driver exists
|
||||
const driverId = 'driver-repo-error-prop';
|
||||
driverRepository.addDriver({
|
||||
id: driverId,
|
||||
name: 'Repo Error Prop Driver',
|
||||
rating: 1000,
|
||||
rank: 1,
|
||||
starts: 0,
|
||||
wins: 0,
|
||||
podiums: 0,
|
||||
leagues: 0,
|
||||
});
|
||||
|
||||
// And: Repository throws error
|
||||
const spy = vi.spyOn(raceRepository, 'getUpcomingRaces').mockRejectedValue(new Error('Repository error'));
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
// Then: Repository error should be propagated
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow('Repository error');
|
||||
|
||||
// And: Error should be catchable by caller
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow('Repository error');
|
||||
|
||||
spy.mockRestore();
|
||||
});
|
||||
});
|
||||
|
||||
describe('Error Logging and Observability', () => {
|
||||
it('should log errors appropriately', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Error logging
|
||||
// Given: A driver exists
|
||||
const driverId = 'driver-logging-error';
|
||||
driverRepository.addDriver({
|
||||
id: driverId,
|
||||
name: 'Logging Error Driver',
|
||||
rating: 1000,
|
||||
rank: 1,
|
||||
starts: 0,
|
||||
wins: 0,
|
||||
podiums: 0,
|
||||
leagues: 0,
|
||||
});
|
||||
|
||||
// And: An error occurs during execution
|
||||
const error = new Error('Logging test error');
|
||||
const spy = vi.spyOn(raceRepository, 'getUpcomingRaces').mockRejectedValue(error);
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
// Then: Error should be logged appropriately
|
||||
// And: Log should include error details
|
||||
// And: Log should include context information
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow('Logging test error');
|
||||
|
||||
// And: Logger should have been called with the error
|
||||
expect(loggerMock.error).toHaveBeenCalledWith(
|
||||
'Failed to fetch dashboard data from repositories',
|
||||
error,
|
||||
expect.objectContaining({ driverId })
|
||||
);
|
||||
|
||||
spy.mockRestore();
|
||||
});
|
||||
|
||||
it('should log event publisher errors', async () => {
|
||||
// Scenario: Event publisher error logging
|
||||
// Given: A driver exists
|
||||
const driverId = 'driver-pub-log-error';
|
||||
driverRepository.addDriver({
|
||||
id: driverId,
|
||||
name: 'Pub Log Error Driver',
|
||||
rating: 1000,
|
||||
rank: 1,
|
||||
starts: 0,
|
||||
wins: 0,
|
||||
podiums: 0,
|
||||
leagues: 0,
|
||||
});
|
||||
|
||||
// And: EventPublisher throws an error
|
||||
const error = new Error('Publisher failed');
|
||||
const spy = vi.spyOn(eventPublisher, 'publishDashboardAccessed').mockRejectedValue(error);
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
await getDashboardUseCase.execute({ driverId });
|
||||
|
||||
// Then: Logger should have been called
|
||||
expect(loggerMock.error).toHaveBeenCalledWith(
|
||||
'Failed to publish dashboard accessed event',
|
||||
error,
|
||||
expect.objectContaining({ driverId })
|
||||
);
|
||||
|
||||
spy.mockRestore();
|
||||
});
|
||||
|
||||
it('should include context in error messages', async () => {
|
||||
// TODO: Implement test
|
||||
// Scenario: Error context
|
||||
// Given: A driver exists
|
||||
const driverId = 'driver-context-error';
|
||||
driverRepository.addDriver({
|
||||
id: driverId,
|
||||
name: 'Context Error Driver',
|
||||
rating: 1000,
|
||||
rank: 1,
|
||||
starts: 0,
|
||||
wins: 0,
|
||||
podiums: 0,
|
||||
leagues: 0,
|
||||
});
|
||||
|
||||
// And: An error occurs during execution
|
||||
const spy = vi.spyOn(raceRepository, 'getUpcomingRaces').mockRejectedValue(new Error('Context test error'));
|
||||
|
||||
// When: GetDashboardUseCase.execute() is called
|
||||
// Then: Error message should include driver ID
|
||||
// And: Error message should include operation details
|
||||
// And: Error message should be informative
|
||||
// Note: The current implementation doesn't include driver ID in error messages
|
||||
// This test documents the expected behavior
|
||||
await expect(getDashboardUseCase.execute({ driverId }))
|
||||
.rejects.toThrow('Context test error');
|
||||
|
||||
// And: EventPublisher should NOT emit any events
|
||||
expect(eventPublisher.getDashboardAccessedEventCount()).toBe(0);
|
||||
|
||||
spy.mockRestore();
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
Reference in New Issue
Block a user