mirror of
https://github.com/crawlab-team/crawlab.git
synced 2026-01-21 17:21:09 +01:00
- Added configuration initialization in db.go to ensure proper setup of application settings. - Refactored runner_test.go to streamline IPC message handling by introducing a setupPipe function and an initRunner function for better readability and maintainability. - Improved synchronization in tests by using channels for signaling readiness and processing, enhancing the reliability of IPC message handling. - Updated test cases to validate IPC message processing and error handling, ensuring robustness in the task runner's functionality.
442 lines
11 KiB
Go
442 lines
11 KiB
Go
package handler
|
|
|
|
import (
|
|
"bufio"
|
|
"context"
|
|
"encoding/json"
|
|
"fmt"
|
|
"io"
|
|
"reflect"
|
|
"runtime"
|
|
"sync"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/crawlab-team/crawlab/core/entity"
|
|
"github.com/crawlab-team/crawlab/core/grpc/client"
|
|
"github.com/crawlab-team/crawlab/core/grpc/server"
|
|
"github.com/crawlab-team/crawlab/core/utils"
|
|
|
|
"github.com/apex/log"
|
|
"github.com/crawlab-team/crawlab/core/constants"
|
|
"github.com/crawlab-team/crawlab/core/models/models"
|
|
"github.com/crawlab-team/crawlab/core/models/service"
|
|
"github.com/spf13/viper"
|
|
"github.com/stretchr/testify/assert"
|
|
"github.com/stretchr/testify/require"
|
|
"go.mongodb.org/mongo-driver/bson/primitive"
|
|
)
|
|
|
|
func setupGrpc(t *testing.T) {
|
|
// Mock IsMaster function by setting viper config
|
|
viper.Set("node.master", true)
|
|
defer viper.Set("node.master", nil) // cleanup after test
|
|
|
|
// Start a gRPC server
|
|
svr := server.GetGrpcServer()
|
|
err := svr.Start()
|
|
require.Nil(t, err)
|
|
|
|
// Start a gRPC client
|
|
client.GetGrpcClient().Start()
|
|
require.Nil(t, err)
|
|
|
|
// Cleanup
|
|
t.Cleanup(func() {
|
|
err = svr.Stop()
|
|
if err != nil {
|
|
log.Warnf("failed to stop gRPC server: %v", err)
|
|
}
|
|
})
|
|
}
|
|
|
|
func setupRunner(t *testing.T) *Runner {
|
|
// Create a test spider
|
|
spider := &models.Spider{
|
|
Name: "Test Spider",
|
|
}
|
|
spiderId, err := service.NewModelService[models.Spider]().InsertOne(*spider)
|
|
require.NoError(t, err)
|
|
spider.Id = spiderId
|
|
|
|
// Create a test task
|
|
task := &models.Task{
|
|
SpiderId: spiderId,
|
|
Status: constants.TaskStatusPending,
|
|
Type: "test",
|
|
Mode: "test",
|
|
NodeId: primitive.NewObjectID(),
|
|
}
|
|
switch runtime.GOOS {
|
|
case "windows":
|
|
task.Cmd = "ping -n 10 127.0.0.1"
|
|
default: // linux and darwin (macOS)
|
|
task.Cmd = "ping -c 10 127.0.0.1"
|
|
}
|
|
taskId, err := service.NewModelService[models.Task]().InsertOne(*task)
|
|
require.NoError(t, err)
|
|
task.Id = taskId
|
|
|
|
// Create a task handler service
|
|
svc := newTaskHandlerService()
|
|
|
|
// Create a task runner
|
|
runner, _ := newTaskRunner(task.Id, svc)
|
|
require.NotNil(t, runner)
|
|
|
|
// Set task and spider
|
|
runner.t = task
|
|
runner.s = spider
|
|
|
|
// Initialize runner
|
|
err = runner.configureCmd()
|
|
require.Nil(t, err)
|
|
|
|
return runner
|
|
}
|
|
|
|
func setupPipe(runner *Runner) (pr *io.PipeReader, pw *io.PipeWriter) {
|
|
// Create a pipe for testing
|
|
pr, pw = io.Pipe()
|
|
runner.stdoutPipe = pr
|
|
runner.cmd.Stdout = pw
|
|
runner.cmd.Stderr = pw
|
|
return pr, pw
|
|
}
|
|
|
|
func initRunner(runner *Runner) chan struct{} {
|
|
// Initialize context and other required fields
|
|
runner.ctx, runner.cancel = context.WithCancel(context.Background())
|
|
runner.wg = sync.WaitGroup{}
|
|
runner.done = make(chan struct{})
|
|
runner.ipcChan = make(chan entity.IPCMessage)
|
|
|
|
// Create a channel to signal that the reader is ready
|
|
readerReady := make(chan struct{})
|
|
|
|
// Start IPC reader with ready signal
|
|
go func() {
|
|
defer runner.wg.Done()
|
|
runner.wg.Add(1)
|
|
close(readerReady) // Signal that reader is ready
|
|
|
|
// Read directly from the pipe for debugging
|
|
scanner := bufio.NewScanner(runner.stdoutPipe)
|
|
for scanner.Scan() {
|
|
line := scanner.Text()
|
|
log.Infof("Read from pipe: %s", line)
|
|
|
|
// Try to parse as IPC message
|
|
var ipcMsg entity.IPCMessage
|
|
if err := json.Unmarshal([]byte(line), &ipcMsg); err != nil {
|
|
log.Errorf("Failed to unmarshal IPC message: %v", err)
|
|
continue
|
|
}
|
|
|
|
if ipcMsg.IPC {
|
|
log.Infof("Valid IPC message received: %+v", ipcMsg)
|
|
if runner.ipcHandler != nil {
|
|
runner.ipcHandler(ipcMsg)
|
|
}
|
|
}
|
|
}
|
|
|
|
if err := scanner.Err(); err != nil {
|
|
log.Errorf("Scanner error: %v", err)
|
|
}
|
|
}()
|
|
|
|
return readerReady
|
|
}
|
|
|
|
func TestRunner(t *testing.T) {
|
|
// Setup test data
|
|
setupGrpc(t)
|
|
|
|
t.Run("HandleIPC", func(t *testing.T) {
|
|
// Create a runner
|
|
runner := setupRunner(t)
|
|
|
|
// Create a pipe for testing
|
|
_, pw := setupPipe(runner)
|
|
|
|
readerReady := initRunner(runner)
|
|
|
|
// Wait for reader to be ready
|
|
<-readerReady
|
|
|
|
// Create test message
|
|
testMsg := entity.IPCMessage{
|
|
Type: "test_type",
|
|
Payload: map[string]interface{}{"key": "value"},
|
|
IPC: true,
|
|
}
|
|
|
|
// Create channels for synchronization
|
|
processed := make(chan bool)
|
|
messageError := make(chan error, 1)
|
|
|
|
// Set up message handler
|
|
runner.SetIPCHandler(func(msg entity.IPCMessage) {
|
|
log.Infof("Handler received IPC message: %+v", msg)
|
|
if msg.Type != testMsg.Type {
|
|
messageError <- fmt.Errorf("expected message type %s, got %s", testMsg.Type, msg.Type)
|
|
return
|
|
}
|
|
if !reflect.DeepEqual(msg.Payload, testMsg.Payload) {
|
|
messageError <- fmt.Errorf("expected payload %v, got %v", testMsg.Payload, msg.Payload)
|
|
return
|
|
}
|
|
processed <- true
|
|
})
|
|
|
|
// Convert message to JSON
|
|
jsonData, err := json.Marshal(testMsg)
|
|
if err != nil {
|
|
t.Fatalf("failed to marshal test message: %v", err)
|
|
}
|
|
|
|
// Write message to pipe
|
|
log.Infof("Writing message to pipe: %s", string(jsonData))
|
|
_, err = fmt.Fprintln(pw, string(jsonData))
|
|
if err != nil {
|
|
t.Fatalf("failed to write to pipe: %v", err)
|
|
}
|
|
log.Info("Message written to pipe")
|
|
|
|
// Wait for message handling with timeout
|
|
select {
|
|
case <-processed:
|
|
log.Info("IPC message was processed successfully")
|
|
case err := <-messageError:
|
|
t.Fatalf("error handling message: %v", err)
|
|
case <-time.After(5 * time.Second):
|
|
t.Fatal("timeout waiting for IPC message to be handled")
|
|
}
|
|
|
|
// Clean up
|
|
runner.cancel() // Cancel context to stop readers
|
|
})
|
|
|
|
t.Run("Cancel", func(t *testing.T) {
|
|
// Create a runner
|
|
runner := setupRunner(t)
|
|
|
|
// Create pipes for stdout
|
|
pr, _ := setupPipe(runner)
|
|
|
|
// Start the command
|
|
err := runner.cmd.Start()
|
|
assert.NoError(t, err)
|
|
log.Infof("started process with PID: %d", runner.cmd.Process.Pid)
|
|
runner.pid = runner.cmd.Process.Pid
|
|
|
|
// Read and print command output
|
|
go func() {
|
|
scanner := bufio.NewScanner(pr)
|
|
for scanner.Scan() {
|
|
log.Info(scanner.Text())
|
|
}
|
|
}()
|
|
|
|
// Wait for process to finish
|
|
go func() {
|
|
err = runner.cmd.Wait()
|
|
if err != nil {
|
|
log.Errorf("process[%d] exited with error: %v", runner.pid, err)
|
|
return
|
|
}
|
|
log.Infof("process[%d] exited successfully", runner.pid)
|
|
}()
|
|
|
|
// Wait for a certain period for the process to start properly
|
|
time.Sleep(1 * time.Second)
|
|
|
|
// Verify process exists before attempting to cancel
|
|
if !utils.ProcessIdExists(runner.pid) {
|
|
require.Fail(t, fmt.Sprintf("Process with PID %d was not started successfully", runner.pid))
|
|
}
|
|
|
|
// Test cancel
|
|
go func() {
|
|
err = runner.Cancel(true)
|
|
assert.NoError(t, err)
|
|
log.Infof("process[%d] cancelled", runner.pid)
|
|
}()
|
|
|
|
// Wait for process to be killed, with shorter timeout
|
|
ctx, cancel := context.WithTimeout(context.Background(), 5*time.Second)
|
|
defer cancel()
|
|
ticker := time.NewTicker(100 * time.Millisecond)
|
|
defer ticker.Stop()
|
|
for {
|
|
select {
|
|
case <-ctx.Done():
|
|
require.Fail(t, fmt.Sprintf("Process with PID %d was not killed within timeout", runner.pid))
|
|
case <-ticker.C:
|
|
exists := utils.ProcessIdExists(runner.pid)
|
|
if !exists {
|
|
return // Exit the test when process is killed
|
|
}
|
|
}
|
|
}
|
|
})
|
|
|
|
t.Run("HandleIPCData", func(t *testing.T) {
|
|
// Create a runner
|
|
runner := setupRunner(t)
|
|
|
|
// Create a pipe for testing
|
|
_, pw := setupPipe(runner)
|
|
|
|
readerReady := initRunner(runner)
|
|
|
|
// Wait for reader to be ready
|
|
<-readerReady
|
|
|
|
// Test cases
|
|
testCases := []struct {
|
|
name string
|
|
message entity.IPCMessage
|
|
expectError bool
|
|
errorTimeout bool
|
|
}{
|
|
{
|
|
name: "valid single object",
|
|
message: entity.IPCMessage{
|
|
Type: constants.IPCMessageData,
|
|
Payload: map[string]interface{}{
|
|
"field1": "value1",
|
|
"field2": 123,
|
|
},
|
|
IPC: true,
|
|
},
|
|
expectError: false,
|
|
},
|
|
{
|
|
name: "valid array of objects",
|
|
message: entity.IPCMessage{
|
|
Type: constants.IPCMessageData,
|
|
Payload: []map[string]interface{}{
|
|
{
|
|
"field1": "value1",
|
|
"field2": 123,
|
|
},
|
|
{
|
|
"field1": "value2",
|
|
"field2": 456,
|
|
},
|
|
},
|
|
IPC: true,
|
|
},
|
|
expectError: false,
|
|
},
|
|
{
|
|
name: "invalid payload type",
|
|
message: entity.IPCMessage{
|
|
Type: constants.IPCMessageData,
|
|
Payload: "invalid",
|
|
IPC: true,
|
|
},
|
|
expectError: true,
|
|
errorTimeout: true,
|
|
},
|
|
{
|
|
name: "non-ipc message",
|
|
message: entity.IPCMessage{
|
|
Type: constants.IPCMessageData,
|
|
Payload: map[string]interface{}{"field": "value"},
|
|
IPC: false,
|
|
},
|
|
expectError: true,
|
|
errorTimeout: true,
|
|
},
|
|
}
|
|
|
|
for _, tc := range testCases {
|
|
t.Run(tc.name, func(t *testing.T) {
|
|
// Create channels for synchronization
|
|
processed := make(chan bool)
|
|
messageError := make(chan error, 1)
|
|
|
|
// Set up message handler
|
|
runner.SetIPCHandler(func(msg entity.IPCMessage) {
|
|
log.Infof("Handler received IPC message: %+v", msg)
|
|
|
|
// Verify message type matches
|
|
if msg.Type != tc.message.Type {
|
|
messageError <- fmt.Errorf("expected message type %s, got %s", tc.message.Type, msg.Type)
|
|
return
|
|
}
|
|
|
|
// Verify IPC flag
|
|
if msg.IPC != tc.message.IPC {
|
|
messageError <- fmt.Errorf("expected IPC flag %v, got %v", tc.message.IPC, msg.IPC)
|
|
return
|
|
}
|
|
|
|
// For data messages, just verify the structure
|
|
if msg.Type == constants.IPCMessageData {
|
|
switch msg.Payload.(type) {
|
|
case map[string]interface{}, []map[string]interface{}, []interface{}:
|
|
processed <- true
|
|
default:
|
|
messageError <- fmt.Errorf("unexpected payload type: %T", msg.Payload)
|
|
}
|
|
return
|
|
}
|
|
|
|
processed <- true
|
|
})
|
|
|
|
// Convert message to JSON
|
|
jsonData, err := json.Marshal(tc.message)
|
|
if err != nil {
|
|
t.Fatalf("failed to marshal test message: %v", err)
|
|
}
|
|
|
|
// Write message to pipe
|
|
log.Infof("Writing message to pipe: %s", string(jsonData))
|
|
_, err = fmt.Fprintln(pw, string(jsonData))
|
|
if err != nil {
|
|
t.Fatalf("failed to write to pipe: %v", err)
|
|
}
|
|
log.Info("Message written to pipe")
|
|
|
|
if tc.expectError {
|
|
if tc.errorTimeout {
|
|
// For invalid messages, expect a timeout
|
|
select {
|
|
case <-processed:
|
|
t.Error("invalid message was unexpectedly processed")
|
|
case <-time.After(1 * time.Second):
|
|
// Success - no processing occurred
|
|
}
|
|
} else {
|
|
// For other error cases, expect an error message
|
|
select {
|
|
case err := <-messageError:
|
|
log.Infof("received expected error: %v", err)
|
|
case <-time.After(5 * time.Second):
|
|
t.Fatal("timeout waiting for error message")
|
|
}
|
|
}
|
|
} else {
|
|
// For valid messages, expect successful processing
|
|
select {
|
|
case <-processed:
|
|
log.Info("IPC message was processed successfully")
|
|
case err := <-messageError:
|
|
t.Fatalf("error handling message: %v", err)
|
|
case <-time.After(5 * time.Second):
|
|
t.Fatal("timeout waiting for IPC message to be handled")
|
|
}
|
|
}
|
|
})
|
|
}
|
|
|
|
// Clean up
|
|
runner.cancel() // Cancel context to stop readers
|
|
})
|
|
}
|