2026-01-08 19:25:52 +01:00
|
|
|
#!/bin/bash
|
|
|
|
|
|
|
|
|
|
# ============================================================================
|
|
|
|
|
# SigNoz Verification Script for Bakery IA
|
|
|
|
|
# ============================================================================
|
|
|
|
|
# This script verifies that SigNoz is properly deployed and functioning
|
|
|
|
|
# ============================================================================
|
|
|
|
|
|
|
|
|
|
set -e
|
|
|
|
|
|
|
|
|
|
# Color codes for output
|
|
|
|
|
RED='\033[0;31m'
|
|
|
|
|
GREEN='\033[0;32m'
|
|
|
|
|
YELLOW='\033[1;33m'
|
|
|
|
|
BLUE='\033[0;34m'
|
|
|
|
|
NC='\033[0m' # No Color
|
|
|
|
|
|
|
|
|
|
# Function to display help
|
|
|
|
|
show_help() {
|
|
|
|
|
echo "Usage: $0 [OPTIONS] ENVIRONMENT"
|
|
|
|
|
echo ""
|
|
|
|
|
echo "Verify SigNoz deployment for Bakery IA"
|
|
|
|
|
echo ""
|
|
|
|
|
echo "Arguments:
|
|
|
|
|
ENVIRONMENT Environment to verify (dev|prod)"
|
|
|
|
|
echo ""
|
|
|
|
|
echo "Options:
|
|
|
|
|
-h, --help Show this help message
|
2026-01-09 07:26:11 +01:00
|
|
|
-n, --namespace NAMESPACE Specify namespace (default: bakery-ia)"
|
2026-01-08 19:25:52 +01:00
|
|
|
echo ""
|
|
|
|
|
echo "Examples:
|
|
|
|
|
$0 dev # Verify development deployment
|
|
|
|
|
$0 prod # Verify production deployment
|
|
|
|
|
$0 --namespace monitoring dev # Verify with custom namespace"
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
# Parse command line arguments
|
2026-01-09 07:26:11 +01:00
|
|
|
NAMESPACE="bakery-ia"
|
2026-01-08 19:25:52 +01:00
|
|
|
|
|
|
|
|
while [[ $# -gt 0 ]]; do
|
|
|
|
|
case $1 in
|
|
|
|
|
-h|--help)
|
|
|
|
|
show_help
|
|
|
|
|
exit 0
|
|
|
|
|
;;
|
|
|
|
|
-n|--namespace)
|
|
|
|
|
NAMESPACE="$2"
|
|
|
|
|
shift 2
|
|
|
|
|
;;
|
|
|
|
|
dev|prod)
|
|
|
|
|
ENVIRONMENT="$1"
|
|
|
|
|
shift
|
|
|
|
|
;;
|
|
|
|
|
*)
|
|
|
|
|
echo "Unknown argument: $1"
|
|
|
|
|
show_help
|
|
|
|
|
exit 1
|
|
|
|
|
;;
|
|
|
|
|
esac
|
|
|
|
|
done
|
|
|
|
|
|
|
|
|
|
# Validate environment
|
|
|
|
|
if [[ -z "$ENVIRONMENT" ]]; then
|
|
|
|
|
echo "Error: Environment not specified. Use 'dev' or 'prod'."
|
|
|
|
|
show_help
|
|
|
|
|
exit 1
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
if [[ "$ENVIRONMENT" != "dev" && "$ENVIRONMENT" != "prod" ]]; then
|
|
|
|
|
echo "Error: Invalid environment. Use 'dev' or 'prod'."
|
|
|
|
|
exit 1
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
# Function to check if kubectl is configured
|
|
|
|
|
check_kubectl() {
|
|
|
|
|
if ! kubectl cluster-info &> /dev/null; then
|
|
|
|
|
echo "${RED}Error: kubectl is not configured or cannot connect to cluster.${NC}"
|
|
|
|
|
echo "Please ensure you have access to a Kubernetes cluster."
|
|
|
|
|
exit 1
|
|
|
|
|
fi
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
# Function to check namespace exists
|
|
|
|
|
check_namespace() {
|
|
|
|
|
if ! kubectl get namespace "$NAMESPACE" &> /dev/null; then
|
|
|
|
|
echo "${RED}Error: Namespace $NAMESPACE does not exist.${NC}"
|
|
|
|
|
echo "Please deploy SigNoz first using: ./deploy-signoz.sh $ENVIRONMENT"
|
|
|
|
|
exit 1
|
|
|
|
|
fi
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
# Function to verify SigNoz deployment
|
|
|
|
|
verify_deployment() {
|
|
|
|
|
echo "${BLUE}"
|
|
|
|
|
echo "=========================================="
|
|
|
|
|
echo "🔍 Verifying SigNoz Deployment"
|
|
|
|
|
echo "=========================================="
|
|
|
|
|
echo "Environment: $ENVIRONMENT"
|
|
|
|
|
echo "Namespace: $NAMESPACE"
|
|
|
|
|
echo "${NC}"
|
|
|
|
|
echo ""
|
|
|
|
|
|
|
|
|
|
# Check if SigNoz helm release exists
|
|
|
|
|
echo "${BLUE}1. Checking Helm release...${NC}"
|
|
|
|
|
if helm list -n "$NAMESPACE" | grep -q signoz; then
|
|
|
|
|
echo "${GREEN}✅ SigNoz Helm release found${NC}"
|
|
|
|
|
else
|
|
|
|
|
echo "${RED}❌ SigNoz Helm release not found${NC}"
|
|
|
|
|
echo "Please deploy SigNoz first using: ./deploy-signoz.sh $ENVIRONMENT"
|
|
|
|
|
exit 1
|
|
|
|
|
fi
|
|
|
|
|
echo ""
|
|
|
|
|
|
|
|
|
|
# Check pod status
|
|
|
|
|
echo "${BLUE}2. Checking pod status...${NC}"
|
|
|
|
|
local total_pods=$(kubectl get pods -n "$NAMESPACE" -l app.kubernetes.io/instance=signoz 2>/dev/null | grep -v "NAME" | wc -l | tr -d ' ' || echo "0")
|
|
|
|
|
local running_pods=$(kubectl get pods -n "$NAMESPACE" -l app.kubernetes.io/instance=signoz --field-selector=status.phase=Running 2>/dev/null | grep -c "Running" || echo "0")
|
|
|
|
|
local ready_pods=$(kubectl get pods -n "$NAMESPACE" -l app.kubernetes.io/instance=signoz 2>/dev/null | grep "Running" | grep "1/1" | wc -l | tr -d ' ' || echo "0")
|
|
|
|
|
|
|
|
|
|
echo "Total pods: $total_pods"
|
|
|
|
|
echo "Running pods: $running_pods"
|
|
|
|
|
echo "Ready pods: $ready_pods"
|
|
|
|
|
|
|
|
|
|
if [[ $total_pods -eq 0 ]]; then
|
|
|
|
|
echo "${RED}❌ No SigNoz pods found${NC}"
|
|
|
|
|
exit 1
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
if [[ $running_pods -eq $total_pods ]]; then
|
|
|
|
|
echo "${GREEN}✅ All pods are running${NC}"
|
|
|
|
|
else
|
|
|
|
|
echo "${YELLOW}⚠️ Some pods are not running${NC}"
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
if [[ $ready_pods -eq $total_pods ]]; then
|
|
|
|
|
echo "${GREEN}✅ All pods are ready${NC}"
|
|
|
|
|
else
|
|
|
|
|
echo "${YELLOW}⚠️ Some pods are not ready${NC}"
|
|
|
|
|
fi
|
|
|
|
|
echo ""
|
|
|
|
|
|
|
|
|
|
# Show pod details
|
|
|
|
|
echo "${BLUE}Pod Details:${NC}"
|
|
|
|
|
kubectl get pods -n "$NAMESPACE" -l app.kubernetes.io/instance=signoz
|
|
|
|
|
echo ""
|
|
|
|
|
|
|
|
|
|
# Check services
|
|
|
|
|
echo "${BLUE}3. Checking services...${NC}"
|
|
|
|
|
local service_count=$(kubectl get svc -n "$NAMESPACE" -l app.kubernetes.io/instance=signoz 2>/dev/null | grep -v "NAME" | wc -l | tr -d ' ' || echo "0")
|
|
|
|
|
|
|
|
|
|
if [[ $service_count -gt 0 ]]; then
|
|
|
|
|
echo "${GREEN}✅ Services found ($service_count services)${NC}"
|
|
|
|
|
kubectl get svc -n "$NAMESPACE" -l app.kubernetes.io/instance=signoz
|
|
|
|
|
else
|
|
|
|
|
echo "${RED}❌ No services found${NC}"
|
|
|
|
|
fi
|
|
|
|
|
echo ""
|
|
|
|
|
|
|
|
|
|
# Check ingress
|
|
|
|
|
echo "${BLUE}4. Checking ingress...${NC}"
|
|
|
|
|
local ingress_count=$(kubectl get ingress -n "$NAMESPACE" -l app.kubernetes.io/instance=signoz 2>/dev/null | grep -v "NAME" | wc -l | tr -d ' ' || echo "0")
|
|
|
|
|
|
|
|
|
|
if [[ $ingress_count -gt 0 ]]; then
|
|
|
|
|
echo "${GREEN}✅ Ingress found ($ingress_count ingress resources)${NC}"
|
|
|
|
|
kubectl get ingress -n "$NAMESPACE" -l app.kubernetes.io/instance=signoz
|
|
|
|
|
else
|
|
|
|
|
echo "${YELLOW}⚠️ No ingress found (may be configured in main namespace)${NC}"
|
|
|
|
|
fi
|
|
|
|
|
echo ""
|
|
|
|
|
|
|
|
|
|
# Check PVCs
|
|
|
|
|
echo "${BLUE}5. Checking persistent volume claims...${NC}"
|
|
|
|
|
local pvc_count=$(kubectl get pvc -n "$NAMESPACE" -l app.kubernetes.io/instance=signoz 2>/dev/null | grep -v "NAME" | wc -l | tr -d ' ' || echo "0")
|
|
|
|
|
|
|
|
|
|
if [[ $pvc_count -gt 0 ]]; then
|
|
|
|
|
echo "${GREEN}✅ PVCs found ($pvc_count PVCs)${NC}"
|
|
|
|
|
kubectl get pvc -n "$NAMESPACE" -l app.kubernetes.io/instance=signoz
|
|
|
|
|
else
|
|
|
|
|
echo "${YELLOW}⚠️ No PVCs found (may not be required for all components)${NC}"
|
|
|
|
|
fi
|
|
|
|
|
echo ""
|
|
|
|
|
|
|
|
|
|
# Check resource usage
|
|
|
|
|
echo "${BLUE}6. Checking resource usage...${NC}"
|
|
|
|
|
if command -v kubectl &> /dev/null && kubectl top pods -n "$NAMESPACE" &> /dev/null; then
|
|
|
|
|
echo "${GREEN}✅ Resource usage:${NC}"
|
|
|
|
|
kubectl top pods -n "$NAMESPACE" -l app.kubernetes.io/instance=signoz
|
|
|
|
|
else
|
|
|
|
|
echo "${YELLOW}⚠️ Metrics server not available or no resource usage data${NC}"
|
|
|
|
|
fi
|
|
|
|
|
echo ""
|
|
|
|
|
|
|
|
|
|
# Check logs for errors
|
|
|
|
|
echo "${BLUE}7. Checking for errors in logs...${NC}"
|
|
|
|
|
local error_found=false
|
|
|
|
|
|
|
|
|
|
# Check each pod for errors
|
|
|
|
|
while IFS= read -r pod; do
|
|
|
|
|
if [[ -n "$pod" ]]; then
|
|
|
|
|
local pod_errors=$(kubectl logs -n "$NAMESPACE" "$pod" 2>/dev/null | grep -i "error\|exception\|fail\|crash" | wc -l || echo "0")
|
|
|
|
|
if [[ $pod_errors -gt 0 ]]; then
|
|
|
|
|
echo "${RED}❌ Errors found in pod $pod ($pod_errors errors)${NC}"
|
|
|
|
|
error_found=true
|
|
|
|
|
fi
|
|
|
|
|
fi
|
|
|
|
|
done < <(kubectl get pods -n "$NAMESPACE" -l app.kubernetes.io/instance=signoz -o name | sed 's|pod/||')
|
|
|
|
|
|
|
|
|
|
if [[ "$error_found" == false ]]; then
|
|
|
|
|
echo "${GREEN}✅ No errors found in logs${NC}"
|
|
|
|
|
fi
|
|
|
|
|
echo ""
|
|
|
|
|
|
|
|
|
|
# Environment-specific checks
|
|
|
|
|
if [[ "$ENVIRONMENT" == "dev" ]]; then
|
|
|
|
|
verify_dev_specific
|
|
|
|
|
else
|
|
|
|
|
verify_prod_specific
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
# Show access information
|
|
|
|
|
show_access_info
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
# Function for development-specific verification
|
|
|
|
|
verify_dev_specific() {
|
|
|
|
|
echo "${BLUE}8. Development-specific checks...${NC}"
|
2026-01-09 07:26:11 +01:00
|
|
|
|
|
|
|
|
# Check if ingress is configured
|
|
|
|
|
if kubectl get ingress -n "$NAMESPACE" 2>/dev/null | grep -q "monitoring.bakery-ia.local"; then
|
|
|
|
|
echo "${GREEN}✅ Development ingress configured${NC}"
|
2026-01-08 19:25:52 +01:00
|
|
|
else
|
2026-01-09 07:26:11 +01:00
|
|
|
echo "${YELLOW}⚠️ Development ingress not found${NC}"
|
2026-01-08 19:25:52 +01:00
|
|
|
fi
|
2026-01-09 07:26:11 +01:00
|
|
|
|
|
|
|
|
# Check unified signoz component resource limits (should be lower for dev)
|
|
|
|
|
local signoz_mem=$(kubectl get deployment -n "$NAMESPACE" -l app.kubernetes.io/component=query-service -o jsonpath='{.items[0].spec.template.spec.containers[0].resources.limits.memory}' 2>/dev/null || echo "")
|
|
|
|
|
if [[ -n "$signoz_mem" ]]; then
|
|
|
|
|
echo "${GREEN}✅ SigNoz component found (memory limit: $signoz_mem)${NC}"
|
|
|
|
|
else
|
|
|
|
|
echo "${YELLOW}⚠️ Could not verify SigNoz component resources${NC}"
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
# Check single replica setup for dev
|
|
|
|
|
local replicas=$(kubectl get deployment -n "$NAMESPACE" -l app.kubernetes.io/component=query-service -o jsonpath='{.items[0].spec.replicas}' 2>/dev/null || echo "0")
|
|
|
|
|
if [[ $replicas -eq 1 ]]; then
|
|
|
|
|
echo "${GREEN}✅ Single replica configuration (appropriate for dev)${NC}"
|
2026-01-08 19:25:52 +01:00
|
|
|
else
|
2026-01-09 07:26:11 +01:00
|
|
|
echo "${YELLOW}⚠️ Multiple replicas detected (replicas: $replicas)${NC}"
|
2026-01-08 19:25:52 +01:00
|
|
|
fi
|
|
|
|
|
echo ""
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
# Function for production-specific verification
|
|
|
|
|
verify_prod_specific() {
|
|
|
|
|
echo "${BLUE}8. Production-specific checks...${NC}"
|
2026-01-09 07:26:11 +01:00
|
|
|
|
2026-01-08 19:25:52 +01:00
|
|
|
# Check if TLS is configured
|
2026-01-09 07:26:11 +01:00
|
|
|
if kubectl get ingress -n "$NAMESPACE" 2>/dev/null | grep -q "signoz-tls"; then
|
2026-01-08 19:25:52 +01:00
|
|
|
echo "${GREEN}✅ TLS certificate configured${NC}"
|
|
|
|
|
else
|
|
|
|
|
echo "${YELLOW}⚠️ TLS certificate not found${NC}"
|
|
|
|
|
fi
|
2026-01-09 07:26:11 +01:00
|
|
|
|
|
|
|
|
# Check if multiple replicas are running for HA
|
|
|
|
|
local signoz_replicas=$(kubectl get deployment -n "$NAMESPACE" -l app.kubernetes.io/component=query-service -o jsonpath='{.items[0].spec.replicas}' 2>/dev/null || echo "1")
|
|
|
|
|
if [[ $signoz_replicas -gt 1 ]]; then
|
|
|
|
|
echo "${GREEN}✅ High availability configured ($signoz_replicas SigNoz replicas)${NC}"
|
2026-01-08 19:25:52 +01:00
|
|
|
else
|
2026-01-09 07:26:11 +01:00
|
|
|
echo "${YELLOW}⚠️ Single SigNoz replica detected (not highly available)${NC}"
|
2026-01-08 19:25:52 +01:00
|
|
|
fi
|
2026-01-09 07:26:11 +01:00
|
|
|
|
|
|
|
|
# Check Zookeeper replicas (critical for production)
|
|
|
|
|
local zk_replicas=$(kubectl get statefulset -n "$NAMESPACE" -l app.kubernetes.io/component=zookeeper -o jsonpath='{.items[0].spec.replicas}' 2>/dev/null || echo "0")
|
|
|
|
|
if [[ $zk_replicas -eq 3 ]]; then
|
|
|
|
|
echo "${GREEN}✅ Zookeeper properly configured with 3 replicas${NC}"
|
|
|
|
|
elif [[ $zk_replicas -gt 0 ]]; then
|
|
|
|
|
echo "${YELLOW}⚠️ Zookeeper has $zk_replicas replicas (recommend 3 for production)${NC}"
|
|
|
|
|
else
|
|
|
|
|
echo "${RED}❌ Zookeeper not found${NC}"
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
# Check OTel Collector replicas
|
|
|
|
|
local otel_replicas=$(kubectl get deployment -n "$NAMESPACE" -l app.kubernetes.io/component=otel-collector -o jsonpath='{.items[0].spec.replicas}' 2>/dev/null || echo "1")
|
|
|
|
|
if [[ $otel_replicas -gt 1 ]]; then
|
|
|
|
|
echo "${GREEN}✅ OTel Collector HA configured ($otel_replicas replicas)${NC}"
|
|
|
|
|
else
|
|
|
|
|
echo "${YELLOW}⚠️ Single OTel Collector replica${NC}"
|
|
|
|
|
fi
|
|
|
|
|
|
2026-01-08 19:25:52 +01:00
|
|
|
# Check resource limits (should be higher for prod)
|
2026-01-09 07:26:11 +01:00
|
|
|
local signoz_mem=$(kubectl get deployment -n "$NAMESPACE" -l app.kubernetes.io/component=query-service -o jsonpath='{.items[0].spec.template.spec.containers[0].resources.limits.memory}' 2>/dev/null || echo "")
|
|
|
|
|
if [[ -n "$signoz_mem" ]]; then
|
|
|
|
|
echo "${GREEN}✅ Production resource limits applied (memory: $signoz_mem)${NC}"
|
2026-01-08 19:25:52 +01:00
|
|
|
else
|
2026-01-09 07:26:11 +01:00
|
|
|
echo "${YELLOW}⚠️ Could not verify resource limits${NC}"
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
# Check HPA (Horizontal Pod Autoscaler)
|
|
|
|
|
local hpa_count=$(kubectl get hpa -n "$NAMESPACE" 2>/dev/null | grep -c signoz || echo "0")
|
|
|
|
|
if [[ $hpa_count -gt 0 ]]; then
|
|
|
|
|
echo "${GREEN}✅ Horizontal Pod Autoscaler configured${NC}"
|
|
|
|
|
else
|
|
|
|
|
echo "${YELLOW}⚠️ No HPA found (consider enabling for production)${NC}"
|
2026-01-08 19:25:52 +01:00
|
|
|
fi
|
|
|
|
|
echo ""
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
# Function to show access information
|
|
|
|
|
show_access_info() {
|
|
|
|
|
echo "${BLUE}"
|
|
|
|
|
echo "=========================================="
|
|
|
|
|
echo "📋 Access Information"
|
|
|
|
|
echo "=========================================="
|
|
|
|
|
echo "${NC}"
|
2026-01-09 07:26:11 +01:00
|
|
|
|
2026-01-08 19:25:52 +01:00
|
|
|
if [[ "$ENVIRONMENT" == "dev" ]]; then
|
2026-01-09 07:26:11 +01:00
|
|
|
echo "SigNoz UI: http://monitoring.bakery-ia.local"
|
|
|
|
|
echo ""
|
|
|
|
|
echo "OpenTelemetry Collector (within cluster):"
|
|
|
|
|
echo " gRPC: signoz-otel-collector.$NAMESPACE.svc.cluster.local:4317"
|
|
|
|
|
echo " HTTP: signoz-otel-collector.$NAMESPACE.svc.cluster.local:4318"
|
2026-01-08 19:25:52 +01:00
|
|
|
echo ""
|
2026-01-09 07:26:11 +01:00
|
|
|
echo "Port-forward for local access:"
|
|
|
|
|
echo " kubectl port-forward -n $NAMESPACE svc/signoz 8080:8080"
|
|
|
|
|
echo " kubectl port-forward -n $NAMESPACE svc/signoz-otel-collector 4317:4317"
|
|
|
|
|
echo " kubectl port-forward -n $NAMESPACE svc/signoz-otel-collector 4318:4318"
|
2026-01-08 19:25:52 +01:00
|
|
|
else
|
2026-01-09 07:26:11 +01:00
|
|
|
echo "SigNoz UI: https://monitoring.bakewise.ai"
|
2026-01-08 19:25:52 +01:00
|
|
|
echo ""
|
2026-01-09 07:26:11 +01:00
|
|
|
echo "OpenTelemetry Collector (within cluster):"
|
|
|
|
|
echo " gRPC: signoz-otel-collector.$NAMESPACE.svc.cluster.local:4317"
|
|
|
|
|
echo " HTTP: signoz-otel-collector.$NAMESPACE.svc.cluster.local:4318"
|
2026-01-08 19:25:52 +01:00
|
|
|
fi
|
2026-01-09 07:26:11 +01:00
|
|
|
|
2026-01-08 19:25:52 +01:00
|
|
|
echo ""
|
|
|
|
|
echo "Default Credentials:"
|
2026-01-09 07:26:11 +01:00
|
|
|
echo " Username: admin@example.com"
|
2026-01-08 19:25:52 +01:00
|
|
|
echo " Password: admin"
|
|
|
|
|
echo ""
|
2026-01-09 07:26:11 +01:00
|
|
|
echo "⚠️ IMPORTANT: Change default password after first login!"
|
|
|
|
|
echo ""
|
|
|
|
|
|
2026-01-08 19:25:52 +01:00
|
|
|
# Show connection test commands
|
|
|
|
|
echo "Connection Test Commands:"
|
|
|
|
|
if [[ "$ENVIRONMENT" == "dev" ]]; then
|
2026-01-09 07:26:11 +01:00
|
|
|
echo " # Test SigNoz UI"
|
|
|
|
|
echo " curl http://monitoring.bakery-ia.local"
|
|
|
|
|
echo ""
|
|
|
|
|
echo " # Test via port-forward"
|
|
|
|
|
echo " kubectl port-forward -n $NAMESPACE svc/signoz 8080:8080"
|
|
|
|
|
echo " curl http://localhost:8080"
|
2026-01-08 19:25:52 +01:00
|
|
|
else
|
2026-01-09 07:26:11 +01:00
|
|
|
echo " # Test SigNoz UI"
|
|
|
|
|
echo " curl https://monitoring.bakewise.ai"
|
|
|
|
|
echo ""
|
|
|
|
|
echo " # Test API health"
|
|
|
|
|
echo " kubectl port-forward -n $NAMESPACE svc/signoz 8080:8080"
|
|
|
|
|
echo " curl http://localhost:8080/api/v1/health"
|
2026-01-08 19:25:52 +01:00
|
|
|
fi
|
|
|
|
|
echo ""
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
# Function to run connectivity tests
|
|
|
|
|
run_connectivity_tests() {
|
|
|
|
|
echo "${BLUE}"
|
|
|
|
|
echo "=========================================="
|
|
|
|
|
echo "🔗 Running Connectivity Tests"
|
|
|
|
|
echo "=========================================="
|
|
|
|
|
echo "${NC}"
|
2026-01-09 07:26:11 +01:00
|
|
|
|
|
|
|
|
# Test pod readiness first
|
|
|
|
|
echo "Checking pod readiness..."
|
|
|
|
|
local ready_pods=$(kubectl get pods -n "$NAMESPACE" -l app.kubernetes.io/instance=signoz --field-selector=status.phase=Running 2>/dev/null | grep "Running" | grep -c "1/1\|2/2" || echo "0")
|
|
|
|
|
local total_pods=$(kubectl get pods -n "$NAMESPACE" -l app.kubernetes.io/instance=signoz 2>/dev/null | grep -v "NAME" | wc -l | tr -d ' ' || echo "0")
|
|
|
|
|
|
|
|
|
|
if [[ $ready_pods -eq $total_pods && $total_pods -gt 0 ]]; then
|
|
|
|
|
echo "${GREEN}✅ All pods are ready ($ready_pods/$total_pods)${NC}"
|
2026-01-08 19:25:52 +01:00
|
|
|
else
|
2026-01-09 07:26:11 +01:00
|
|
|
echo "${YELLOW}⚠️ Some pods not ready ($ready_pods/$total_pods)${NC}"
|
|
|
|
|
fi
|
|
|
|
|
echo ""
|
|
|
|
|
|
|
|
|
|
# Test internal service connectivity
|
|
|
|
|
echo "Testing internal service connectivity..."
|
|
|
|
|
local signoz_svc=$(kubectl get svc -n "$NAMESPACE" signoz -o jsonpath='{.spec.clusterIP}' 2>/dev/null || echo "")
|
|
|
|
|
if [[ -n "$signoz_svc" ]]; then
|
|
|
|
|
echo "${GREEN}✅ SigNoz service accessible at $signoz_svc:8080${NC}"
|
|
|
|
|
else
|
|
|
|
|
echo "${RED}❌ SigNoz service not found${NC}"
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
local otel_svc=$(kubectl get svc -n "$NAMESPACE" signoz-otel-collector -o jsonpath='{.spec.clusterIP}' 2>/dev/null || echo "")
|
|
|
|
|
if [[ -n "$otel_svc" ]]; then
|
|
|
|
|
echo "${GREEN}✅ OTel Collector service accessible at $otel_svc:4317 (gRPC), $otel_svc:4318 (HTTP)${NC}"
|
|
|
|
|
else
|
|
|
|
|
echo "${RED}❌ OTel Collector service not found${NC}"
|
|
|
|
|
fi
|
|
|
|
|
echo ""
|
|
|
|
|
|
|
|
|
|
if [[ "$ENVIRONMENT" == "prod" ]]; then
|
2026-01-08 19:25:52 +01:00
|
|
|
echo "${YELLOW}⚠️ Production connectivity tests require valid DNS and TLS${NC}"
|
|
|
|
|
echo " Please ensure monitoring.bakewise.ai resolves to your cluster"
|
2026-01-09 07:26:11 +01:00
|
|
|
echo ""
|
|
|
|
|
echo "Manual test:"
|
|
|
|
|
echo " curl -I https://monitoring.bakewise.ai"
|
2026-01-08 19:25:52 +01:00
|
|
|
fi
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
# Main execution
|
|
|
|
|
main() {
|
|
|
|
|
echo "${BLUE}"
|
|
|
|
|
echo "=========================================="
|
|
|
|
|
echo "🔍 SigNoz Verification for Bakery IA"
|
|
|
|
|
echo "=========================================="
|
|
|
|
|
echo "${NC}"
|
|
|
|
|
|
|
|
|
|
# Check prerequisites
|
|
|
|
|
check_kubectl
|
|
|
|
|
check_namespace
|
|
|
|
|
|
|
|
|
|
# Verify deployment
|
|
|
|
|
verify_deployment
|
|
|
|
|
|
|
|
|
|
# Run connectivity tests
|
|
|
|
|
run_connectivity_tests
|
|
|
|
|
|
|
|
|
|
echo "${GREEN}"
|
|
|
|
|
echo "=========================================="
|
|
|
|
|
echo "✅ Verification Complete"
|
|
|
|
|
echo "=========================================="
|
|
|
|
|
echo "${NC}"
|
|
|
|
|
|
|
|
|
|
echo "Summary:"
|
|
|
|
|
echo " Environment: $ENVIRONMENT"
|
|
|
|
|
echo " Namespace: $NAMESPACE"
|
|
|
|
|
echo ""
|
|
|
|
|
echo "Next Steps:"
|
|
|
|
|
echo " 1. Access SigNoz UI and verify dashboards"
|
|
|
|
|
echo " 2. Configure alert rules for your services"
|
|
|
|
|
echo " 3. Instrument your applications with OpenTelemetry"
|
|
|
|
|
echo " 4. Set up custom dashboards for key metrics"
|
|
|
|
|
echo ""
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
# Run main function
|
|
|
|
|
main
|