compiles again
This commit is contained in:
@ -1,6 +1,6 @@
|
||||
#include "CatchFactory.h"
|
||||
|
||||
#include <fsfw/datapoollocal/LocalDataPoolManager.h>
|
||||
// #include <fsfw/datapoollocal/LocalDataPoolManager.h>
|
||||
#include <fsfw/devicehandlers/DeviceHandlerBase.h>
|
||||
#include <fsfw/events/EventManager.h>
|
||||
#include <fsfw/health/HealthTable.h>
|
||||
@ -57,7 +57,7 @@ void Factory::setStaticFrameworkObjectIds() {
|
||||
DeviceHandlerBase::powerSwitcherId = objects::NO_OBJECT;
|
||||
DeviceHandlerBase::rawDataReceiverId = objects::NO_OBJECT;
|
||||
|
||||
PeriodicHkGenerationHelper::defaultHkDestination = objects::NO_OBJECT;
|
||||
// PeriodicHkGenerationHelper::defaultHkDestination = objects::NO_OBJECT;
|
||||
DeviceHandlerFailureIsolation::powerConfirmationId = objects::NO_OBJECT;
|
||||
}
|
||||
|
||||
|
@ -1,5 +1,5 @@
|
||||
#include <fsfw/datapool/PoolReadGuard.h>
|
||||
#include <fsfw/datapoollocal/HasLocalDataPoolIF.h>
|
||||
// #include <fsfw/datapoollocal/HasLocalDataPoolIF.h>
|
||||
#include <fsfw/datapoollocal/SharedLocalDataSet.h>
|
||||
#include <fsfw/datapoollocal/StaticLocalDataSet.h>
|
||||
#include <fsfw/globalfunctions/bitutility.h>
|
||||
@ -18,8 +18,7 @@ using namespace returnvalue;
|
||||
TEST_CASE("DataSetTest", "[DataSetTest]") {
|
||||
auto queue = MessageQueueMock(1);
|
||||
LocalPoolOwnerBase poolOwner(queue, objects::TEST_LOCAL_POOL_OWNER_BASE);
|
||||
REQUIRE(poolOwner.initializeHkManager() == OK);
|
||||
REQUIRE(poolOwner.initializeHkManagerAfterTaskCreation() == OK);
|
||||
poolOwner.initialize();
|
||||
LocalPoolStaticTestDataSet localSet;
|
||||
|
||||
SECTION("BasicTest") {
|
||||
@ -59,15 +58,15 @@ TEST_CASE("DataSetTest", "[DataSetTest]") {
|
||||
/* Test read operation. Values should be all zeros */
|
||||
PoolReadGuard readHelper(&localSet);
|
||||
REQUIRE(readHelper.getReadResult() == returnvalue::OK);
|
||||
CHECK(not localSet.isValid());
|
||||
// CHECK(not localSet.isValid());
|
||||
CHECK(localSet.localPoolVarUint8.value == 0);
|
||||
CHECK(not localSet.localPoolVarUint8.isValid());
|
||||
// CHECK(not localSet.localPoolVarUint8.isValid());
|
||||
CHECK(localSet.localPoolVarFloat.value == Catch::Approx(0.0));
|
||||
CHECK(not localSet.localPoolVarUint8.isValid());
|
||||
// CHECK(not localSet.localPoolVarUint8.isValid());
|
||||
CHECK(localSet.localPoolUint16Vec.value[0] == 0);
|
||||
CHECK(localSet.localPoolUint16Vec.value[1] == 0);
|
||||
CHECK(localSet.localPoolUint16Vec.value[2] == 0);
|
||||
CHECK(not localSet.localPoolVarUint8.isValid());
|
||||
// CHECK(not localSet.localPoolVarUint8.isValid());
|
||||
|
||||
/* Now set new values, commit should be done by read helper automatically */
|
||||
localSet.localPoolVarUint8 = 232;
|
||||
@ -75,7 +74,7 @@ TEST_CASE("DataSetTest", "[DataSetTest]") {
|
||||
localSet.localPoolUint16Vec.value[0] = 232;
|
||||
localSet.localPoolUint16Vec.value[1] = 23923;
|
||||
localSet.localPoolUint16Vec.value[2] = 1;
|
||||
localSet.setValidity(true, true);
|
||||
// localSet.setValidity(true, true);
|
||||
}
|
||||
|
||||
/* Zero out some values for next test */
|
||||
@ -92,18 +91,17 @@ TEST_CASE("DataSetTest", "[DataSetTest]") {
|
||||
the values in the pool */
|
||||
PoolReadGuard readHelper(&localSet);
|
||||
REQUIRE(readHelper.getReadResult() == returnvalue::OK);
|
||||
CHECK(localSet.isValid());
|
||||
// CHECK(localSet.isValid());
|
||||
CHECK(localSet.localPoolVarUint8.value == 232);
|
||||
CHECK(localSet.localPoolVarUint8.isValid());
|
||||
// CHECK(localSet.localPoolVarUint8.isValid());
|
||||
CHECK(localSet.localPoolVarFloat.value == Catch::Approx(-2324.322));
|
||||
CHECK(localSet.localPoolVarFloat.isValid());
|
||||
// CHECK(localSet.localPoolVarFloat.isValid());
|
||||
CHECK(localSet.localPoolUint16Vec.value[0] == 232);
|
||||
CHECK(localSet.localPoolUint16Vec.value[1] == 23923);
|
||||
CHECK(localSet.localPoolUint16Vec.value[2] == 1);
|
||||
CHECK(localSet.localPoolUint16Vec.isValid());
|
||||
// CHECK(localSet.localPoolUint16Vec.isValid());
|
||||
|
||||
/* Now we serialize these values into a buffer without the validity buffer */
|
||||
localSet.setValidityBufferGeneration(false);
|
||||
// Now we serialize these values into a buffer without the validity buffer
|
||||
maxSize = localSet.getSerializedSize();
|
||||
CHECK(maxSize == sizeof(uint8_t) + sizeof(uint16_t) * 3 + sizeof(float));
|
||||
serSize = 0;
|
||||
@ -138,9 +136,9 @@ TEST_CASE("DataSetTest", "[DataSetTest]") {
|
||||
CHECK(localSet.localPoolUint16Vec.value[1] == 0);
|
||||
CHECK(localSet.localPoolUint16Vec.value[2] == 0);
|
||||
/* Validity should be unchanged */
|
||||
CHECK(localSet.localPoolVarUint8.isValid());
|
||||
CHECK(localSet.localPoolVarFloat.isValid());
|
||||
CHECK(localSet.localPoolUint16Vec.isValid());
|
||||
// CHECK(localSet.localPoolVarUint8.isValid());
|
||||
// CHECK(localSet.localPoolVarFloat.isValid());
|
||||
// CHECK(localSet.localPoolUint16Vec.isValid());
|
||||
|
||||
/* Now we do the same process but with the validity buffer */
|
||||
localSet.localPoolVarUint8 = 232;
|
||||
@ -148,10 +146,10 @@ TEST_CASE("DataSetTest", "[DataSetTest]") {
|
||||
localSet.localPoolUint16Vec.value[0] = 232;
|
||||
localSet.localPoolUint16Vec.value[1] = 23923;
|
||||
localSet.localPoolUint16Vec.value[2] = 1;
|
||||
localSet.localPoolVarUint8.setValid(true);
|
||||
localSet.localPoolVarFloat.setValid(false);
|
||||
localSet.localPoolUint16Vec.setValid(true);
|
||||
localSet.setValidityBufferGeneration(true);
|
||||
// localSet.localPoolVarUint8.setValid(true);
|
||||
// localSet.localPoolVarFloat.setValid(false);
|
||||
// localSet.localPoolUint16Vec.setValid(true);
|
||||
// localSet.setValidityBufferGeneration(true);
|
||||
maxSize = localSet.getSerializedSize();
|
||||
CHECK(maxSize == sizeof(uint8_t) + sizeof(uint16_t) * 3 + sizeof(float) + 1);
|
||||
serSize = 0;
|
||||
@ -195,9 +193,9 @@ TEST_CASE("DataSetTest", "[DataSetTest]") {
|
||||
CHECK(localSet.localPoolUint16Vec.value[0] == 0);
|
||||
CHECK(localSet.localPoolUint16Vec.value[1] == 0);
|
||||
CHECK(localSet.localPoolUint16Vec.value[2] == 0);
|
||||
CHECK(not localSet.localPoolVarUint8.isValid());
|
||||
CHECK(localSet.localPoolVarFloat.isValid());
|
||||
CHECK(not localSet.localPoolUint16Vec.isValid());
|
||||
// CHECK(not localSet.localPoolVarUint8.isValid());
|
||||
// CHECK(localSet.localPoolVarFloat.isValid());
|
||||
// CHECK(not localSet.localPoolUint16Vec.isValid());
|
||||
}
|
||||
|
||||
/* Common fault test cases */
|
||||
@ -210,7 +208,7 @@ TEST_CASE("DataSetTest", "[DataSetTest]") {
|
||||
}
|
||||
|
||||
SECTION("MorePoolVariables") {
|
||||
LocalDataSet set(&poolOwner, 2, 10);
|
||||
LocalDataSet set(poolOwner.sharedPool, 2, 10);
|
||||
|
||||
/* Register same variables again to get more than 8 registered variables */
|
||||
for (uint8_t idx = 0; idx < 8; idx++) {
|
||||
@ -219,12 +217,12 @@ TEST_CASE("DataSetTest", "[DataSetTest]") {
|
||||
REQUIRE(set.registerVariable(&localSet.localPoolVarUint8) == returnvalue::OK);
|
||||
REQUIRE(set.registerVariable(&localSet.localPoolUint16Vec) == returnvalue::OK);
|
||||
|
||||
set.setValidityBufferGeneration(true);
|
||||
// set.setValidityBufferGeneration(true);
|
||||
{
|
||||
PoolReadGuard readHelper(&localSet);
|
||||
localSet.localPoolVarUint8.value = 42;
|
||||
localSet.localPoolVarUint8.setValid(true);
|
||||
localSet.localPoolUint16Vec.setValid(false);
|
||||
// localSet.localPoolVarUint8.setValid(true);
|
||||
// localSet.localPoolUint16Vec.setValid(false);
|
||||
}
|
||||
|
||||
size_t maxSize = set.getSerializedSize();
|
||||
@ -252,13 +250,13 @@ TEST_CASE("DataSetTest", "[DataSetTest]") {
|
||||
size_t sizeToDeSerialize = serSize;
|
||||
CHECK(set.deSerialize(&constBuffPtr, &sizeToDeSerialize, SerializeIF::Endianness::MACHINE) ==
|
||||
returnvalue::OK);
|
||||
CHECK(localSet.localPoolVarUint8.isValid() == false);
|
||||
CHECK(localSet.localPoolUint16Vec.isValid() == true);
|
||||
// CHECK(localSet.localPoolVarUint8.isValid() == false);
|
||||
// CHECK(localSet.localPoolUint16Vec.isValid() == true);
|
||||
}
|
||||
|
||||
SECTION("SharedDataSet") {
|
||||
object_id_t sharedSetId = objects::SHARED_SET_ID;
|
||||
SharedLocalDataSet sharedSet(sharedSetId, &poolOwner, lpool::testSetId, 5);
|
||||
SharedLocalDataSet sharedSet(sharedSetId, poolOwner.sharedPool, lpool::testSetId, 5);
|
||||
localSet.localPoolVarUint8.setReadWriteMode(pool_rwm_t::VAR_WRITE);
|
||||
localSet.localPoolUint16Vec.setReadWriteMode(pool_rwm_t::VAR_WRITE);
|
||||
CHECK(sharedSet.registerVariable(&localSet.localPoolVarUint8) == returnvalue::OK);
|
||||
@ -277,6 +275,6 @@ TEST_CASE("DataSetTest", "[DataSetTest]") {
|
||||
CHECK(sharedSet.commit() == returnvalue::OK);
|
||||
}
|
||||
|
||||
sharedSet.setReadCommitProtectionBehaviour(true);
|
||||
// sharedSet.setReadCommitProtectionBehaviour(true);
|
||||
}
|
||||
}
|
||||
|
@ -1,5 +1,5 @@
|
||||
#include <fsfw/datapool/PoolReadGuard.h>
|
||||
#include <fsfw/datapoollocal/HasLocalDataPoolIF.h>
|
||||
#include <fsfw/datapoollocal/PeriodicHkGenerationHelper.h>
|
||||
#include <fsfw/datapoollocal/StaticLocalDataSet.h>
|
||||
#include <fsfw/globalfunctions/timevalOperations.h>
|
||||
#include <fsfw/housekeeping/HousekeepingSnapshot.h>
|
||||
@ -23,8 +23,8 @@ TEST_CASE("Local Pool Manager Tests", "[LocManTest]") {
|
||||
auto hkReceiver = HkReceiverMock(hkDest);
|
||||
auto queue = MessageQueueMock(3);
|
||||
LocalPoolOwnerBase poolOwner(queue, objects::TEST_LOCAL_POOL_OWNER_BASE);
|
||||
REQUIRE(poolOwner.initializeHkManager() == returnvalue::OK);
|
||||
REQUIRE(poolOwner.initializeHkManagerAfterTaskCreation() == returnvalue::OK);
|
||||
|
||||
REQUIRE(poolOwner.initialize() == returnvalue::OK);
|
||||
|
||||
MessageQueueMock& poolOwnerMock = poolOwner.getMockQueueHandle();
|
||||
|
||||
@ -33,37 +33,37 @@ TEST_CASE("Local Pool Manager Tests", "[LocManTest]") {
|
||||
poolOwnerMock.setDefaultDestination(defaultDestId);
|
||||
poolOwner.setHkDestId(hkDest);
|
||||
|
||||
auto* hkMan = poolOwner.getHkManagerHandle();
|
||||
|
||||
CommandMessage messageSent;
|
||||
|
||||
// TODO: Fix
|
||||
/*
|
||||
SECTION("Basic Test") {
|
||||
{
|
||||
/* For code coverage, should not crash */
|
||||
// For code coverage, should not crash
|
||||
PeriodicHkGenerationHelper manager(nullptr, nullptr);
|
||||
}
|
||||
auto owner = poolOwner.poolManager.getOwner();
|
||||
auto owner = poolOwner.hkHelper.getOwner();
|
||||
REQUIRE(owner != nullptr);
|
||||
CHECK(owner->getObjectId() == objects::TEST_LOCAL_POOL_OWNER_BASE);
|
||||
|
||||
/* Subscribe for message generation on update. */
|
||||
// Subscribe for message generation on update.
|
||||
REQUIRE(poolOwner.subscribeWrapperSetUpdate(subscriberId) == returnvalue::OK);
|
||||
/* Subscribe for an update message. */
|
||||
/// Subscribe for an update message.
|
||||
poolOwner.dataset.setChanged(true);
|
||||
/* Now the update message should be generated. */
|
||||
REQUIRE(poolOwner.poolManager.performHkOperation() == returnvalue::OK);
|
||||
// Now the update message should be generated.
|
||||
REQUIRE(poolOwner.hkHelper.performHkOperation() == returnvalue::OK);
|
||||
REQUIRE(poolOwnerMock.wasMessageSent());
|
||||
|
||||
REQUIRE(poolOwnerMock.getNextSentMessage(subscriberId, messageSent) == returnvalue::OK);
|
||||
CHECK(messageSent.getCommand() ==
|
||||
static_cast<int>(HousekeepingMessage::UPDATE_NOTIFICATION_SET));
|
||||
|
||||
/* Should have been reset. */
|
||||
// Should have been reset.
|
||||
CHECK(poolOwner.dataset.hasChanged() == false);
|
||||
poolOwnerMock.clearMessages(true);
|
||||
/* Set changed again, result should be the same. */
|
||||
// Set changed again, result should be the same.
|
||||
poolOwner.dataset.setChanged(true);
|
||||
REQUIRE(poolOwner.poolManager.performHkOperation() == returnvalue::OK);
|
||||
REQUIRE(poolOwner.hkHelper.performHkOperation() == returnvalue::OK);
|
||||
|
||||
REQUIRE(poolOwnerMock.wasMessageSent() == true);
|
||||
CHECK(poolOwnerMock.numberOfSentMessages() == 1);
|
||||
@ -72,10 +72,10 @@ TEST_CASE("Local Pool Manager Tests", "[LocManTest]") {
|
||||
static_cast<int>(HousekeepingMessage::UPDATE_NOTIFICATION_SET));
|
||||
|
||||
poolOwnerMock.clearMessages(true);
|
||||
/* Now subscribe for set update HK as well. */
|
||||
// Now subscribe for set update HK as well.
|
||||
REQUIRE(poolOwner.subscribeWrapperSetUpdateHk(false, &hkReceiver) == returnvalue::OK);
|
||||
poolOwner.dataset.setChanged(true);
|
||||
REQUIRE(poolOwner.poolManager.performHkOperation() == returnvalue::OK);
|
||||
REQUIRE(poolOwner.hkHelper.performHkOperation() == returnvalue::OK);
|
||||
REQUIRE(poolOwnerMock.wasMessageSent() == true);
|
||||
CHECK(poolOwnerMock.numberOfSentMessages() == 2);
|
||||
// first message sent should be the update notification
|
||||
@ -84,12 +84,12 @@ TEST_CASE("Local Pool Manager Tests", "[LocManTest]") {
|
||||
static_cast<int>(HousekeepingMessage::UPDATE_NOTIFICATION_SET));
|
||||
REQUIRE(poolOwnerMock.getNextSentMessageToDefaultDest(messageSent) == returnvalue::OK);
|
||||
CHECK(messageSent.getCommand() == static_cast<int>(HousekeepingMessage::HK_REPORT));
|
||||
/* Clear message to avoid memory leak, our mock won't do it for us (yet) */
|
||||
// Clear message to avoid memory leak, our mock won't do it for us (yet)
|
||||
CommandMessageCleaner::clearCommandMessage(&messageSent);
|
||||
}
|
||||
|
||||
SECTION("SetSnapshotUpdateTest") {
|
||||
/* Set the variables in the set to certain values. These are checked later. */
|
||||
// Set the variables in the set to certain values. These are checked later.
|
||||
{
|
||||
PoolReadGuard readHelper(&poolOwner.dataset);
|
||||
REQUIRE(readHelper.getReadResult() == returnvalue::OK);
|
||||
@ -100,22 +100,22 @@ TEST_CASE("Local Pool Manager Tests", "[LocManTest]") {
|
||||
poolOwner.dataset.localPoolUint16Vec.value[2] = 42932;
|
||||
}
|
||||
|
||||
/* Subscribe for snapshot generation on update. */
|
||||
// Subscribe for snapshot generation on update.
|
||||
REQUIRE(poolOwner.subscribeWrapperSetUpdateSnapshot(subscriberId) == returnvalue::OK);
|
||||
poolOwner.dataset.setChanged(true);
|
||||
|
||||
/* Store current time, we are going to check the (approximate) time equality later */
|
||||
// Store current time, we are going to check the (approximate) time equality later
|
||||
timeval now{};
|
||||
Clock::getClock_timeval(&now);
|
||||
|
||||
/* Trigger generation of snapshot */
|
||||
REQUIRE(poolOwner.poolManager.performHkOperation() == returnvalue::OK);
|
||||
// Trigger generation of snapshot
|
||||
REQUIRE(poolOwner.hkHelper.performHkOperation() == returnvalue::OK);
|
||||
REQUIRE(poolOwnerMock.wasMessageSent());
|
||||
CHECK(poolOwnerMock.numberOfSentMessages() == 1);
|
||||
REQUIRE(poolOwnerMock.getNextSentMessage(subscriberId, messageSent) == returnvalue::OK);
|
||||
/* Check that snapshot was generated */
|
||||
// Check that snapshot was generated
|
||||
CHECK(messageSent.getCommand() == static_cast<int>(HousekeepingMessage::UPDATE_SNAPSHOT_SET));
|
||||
/* Now we deserialize the snapshot into a new dataset instance */
|
||||
// Now we deserialize the snapshot into a new dataset instance
|
||||
CCSDSTime::CDS_short cdsShort{};
|
||||
LocalPoolTestDataSet newSet;
|
||||
HousekeepingSnapshot snapshot(&cdsShort, &newSet);
|
||||
@ -130,17 +130,17 @@ TEST_CASE("Local Pool Manager Tests", "[LocManTest]") {
|
||||
CHECK(newSet.localPoolUint16Vec.value[0] == 0);
|
||||
CHECK(newSet.localPoolUint16Vec.value[1] == 0);
|
||||
CHECK(newSet.localPoolUint16Vec.value[2] == 0);
|
||||
/* Fill the dataset and timestamp */
|
||||
// Fill the dataset and timestamp
|
||||
REQUIRE(snapshot.deSerialize(&readOnlyPtr, &sizeToDeserialize,
|
||||
SerializeIF::Endianness::MACHINE) == returnvalue::OK);
|
||||
/* Now we check that the snapshot is actually correct */
|
||||
// Now we check that the snapshot is actually correct
|
||||
CHECK(newSet.localPoolVarFloat.value == Catch::Approx(-12.242));
|
||||
CHECK(newSet.localPoolVarUint8 == 5);
|
||||
CHECK(newSet.localPoolUint16Vec.value[0] == 2);
|
||||
CHECK(newSet.localPoolUint16Vec.value[1] == 32);
|
||||
CHECK(newSet.localPoolUint16Vec.value[2] == 42932);
|
||||
|
||||
/* Now we check that both times are equal */
|
||||
// Now we check that both times are equal
|
||||
timeval timeFromHK{};
|
||||
auto result = CCSDSTime::convertFromCDS(&timeFromHK, &cdsShort);
|
||||
CHECK(result == returnvalue::OK);
|
||||
@ -149,11 +149,11 @@ TEST_CASE("Local Pool Manager Tests", "[LocManTest]") {
|
||||
}
|
||||
|
||||
SECTION("VariableSnapshotTest") {
|
||||
/* Acquire subscription interface */
|
||||
// Acquire subscription interface
|
||||
PeriodicHkGenerationProviderIF* subscriptionIF = poolOwner.getSubscriptionInterface();
|
||||
REQUIRE(subscriptionIF != nullptr);
|
||||
|
||||
/* Subscribe for variable snapshot */
|
||||
// Subscribe for variable snapshot
|
||||
REQUIRE(poolOwner.subscribeWrapperVariableSnapshot(subscriberId, lpool::uint8VarId) ==
|
||||
returnvalue::OK);
|
||||
auto poolVar =
|
||||
@ -167,22 +167,22 @@ TEST_CASE("Local Pool Manager Tests", "[LocManTest]") {
|
||||
}
|
||||
|
||||
poolVar->setChanged(true);
|
||||
/* Store current time, we are going to check the (approximate) time equality later */
|
||||
// Store current time, we are going to check the (approximate) time equality later
|
||||
CCSDSTime::CDS_short timeCdsNow{};
|
||||
timeval now{};
|
||||
Clock::getClock_timeval(&now);
|
||||
REQUIRE(poolOwner.poolManager.performHkOperation() == returnvalue::OK);
|
||||
REQUIRE(poolOwner.hkHelper.performHkOperation() == returnvalue::OK);
|
||||
|
||||
/* Check update snapshot was sent. */
|
||||
// Check update snapshot was sent.
|
||||
REQUIRE(poolOwnerMock.wasMessageSent());
|
||||
CHECK(poolOwnerMock.numberOfSentMessages() == 1);
|
||||
|
||||
/* Should have been reset. */
|
||||
// Should have been reset.
|
||||
CHECK(poolVar->hasChanged() == false);
|
||||
REQUIRE(poolOwnerMock.getNextSentMessage(subscriberId, messageSent) == returnvalue::OK);
|
||||
CHECK(messageSent.getCommand() ==
|
||||
static_cast<int>(HousekeepingMessage::UPDATE_SNAPSHOT_VARIABLE));
|
||||
/* Now we deserialize the snapshot into a new dataset instance */
|
||||
// Now we deserialize the snapshot into a new dataset instance
|
||||
CCSDSTime::CDS_short cdsShort{};
|
||||
lp_var_t<uint8_t> varCopy = lp_var_t<uint8_t>(lpool::uint8VarGpid);
|
||||
HousekeepingSnapshot snapshot(&cdsShort, &varCopy);
|
||||
@ -193,12 +193,12 @@ TEST_CASE("Local Pool Manager Tests", "[LocManTest]") {
|
||||
const uint8_t* readOnlyPtr = accessorPair.second.data();
|
||||
size_t sizeToDeserialize = accessorPair.second.size();
|
||||
CHECK(varCopy.value == 0);
|
||||
/* Fill the dataset and timestamp */
|
||||
// Fill the dataset and timestamp
|
||||
REQUIRE(snapshot.deSerialize(&readOnlyPtr, &sizeToDeserialize,
|
||||
SerializeIF::Endianness::MACHINE) == returnvalue::OK);
|
||||
CHECK(varCopy.value == 25);
|
||||
|
||||
/* Now we check that both times are equal */
|
||||
// Now we check that both times are equal
|
||||
timeval timeFromHK{};
|
||||
auto result = CCSDSTime::convertFromCDS(&timeFromHK, &cdsShort);
|
||||
CHECK(result == returnvalue::OK);
|
||||
@ -207,11 +207,11 @@ TEST_CASE("Local Pool Manager Tests", "[LocManTest]") {
|
||||
}
|
||||
|
||||
SECTION("VariableNotificationTest") {
|
||||
/* Acquire subscription interface */
|
||||
// Acquire subscription interface
|
||||
PeriodicHkGenerationProviderIF* subscriptionIF = poolOwner.getSubscriptionInterface();
|
||||
REQUIRE(subscriptionIF != nullptr);
|
||||
|
||||
/* Subscribe for variable update */
|
||||
// Subscribe for variable update
|
||||
REQUIRE(poolOwner.subscribeWrapperVariableUpdate(subscriberId, lpool::uint8VarId) ==
|
||||
returnvalue::OK);
|
||||
auto* poolVar =
|
||||
@ -219,25 +219,25 @@ TEST_CASE("Local Pool Manager Tests", "[LocManTest]") {
|
||||
REQUIRE(poolVar != nullptr);
|
||||
poolVar->setChanged(true);
|
||||
REQUIRE(poolVar->hasChanged() == true);
|
||||
REQUIRE(poolOwner.poolManager.performHkOperation() == returnvalue::OK);
|
||||
REQUIRE(poolOwner.hkHelper.performHkOperation() == returnvalue::OK);
|
||||
|
||||
/* Check update notification was sent. */
|
||||
// Check update notification was sent.
|
||||
REQUIRE(poolOwnerMock.wasMessageSent());
|
||||
CHECK(poolOwnerMock.numberOfSentMessages() == 1);
|
||||
/* Should have been reset. */
|
||||
// Should have been reset.
|
||||
CHECK(poolVar->hasChanged() == false);
|
||||
REQUIRE(poolOwnerMock.getNextSentMessage(subscriberId, messageSent) == returnvalue::OK);
|
||||
CHECK(messageSent.getCommand() ==
|
||||
static_cast<int>(HousekeepingMessage::UPDATE_NOTIFICATION_VARIABLE));
|
||||
/* Now subscribe for the dataset update (HK and update) again with subscription interface */
|
||||
// Now subscribe for the dataset update (HK and update) again with subscription interface
|
||||
REQUIRE(subscriptionIF->subscribeForSetUpdateMessage(lpool::testSetId, objects::NO_OBJECT,
|
||||
subscriberId, false) == returnvalue::OK);
|
||||
REQUIRE(poolOwner.subscribeWrapperSetUpdateHk(false, &hkReceiver) == returnvalue::OK);
|
||||
|
||||
poolOwner.dataset.setChanged(true);
|
||||
poolOwnerMock.clearMessages();
|
||||
REQUIRE(poolOwner.poolManager.performHkOperation() == returnvalue::OK);
|
||||
/* Now two messages should be sent. */
|
||||
REQUIRE(poolOwner.hkHelper.performHkOperation() == returnvalue::OK);
|
||||
// Now two messages should be sent.
|
||||
REQUIRE(poolOwnerMock.wasMessageSent());
|
||||
CHECK(poolOwnerMock.numberOfSentMessages() == 2);
|
||||
poolOwnerMock.clearMessages(true);
|
||||
@ -245,8 +245,8 @@ TEST_CASE("Local Pool Manager Tests", "[LocManTest]") {
|
||||
poolOwner.dataset.setChanged(true);
|
||||
poolOwnerMock.clearMessages(true);
|
||||
poolVar->setChanged(true);
|
||||
REQUIRE(poolOwner.poolManager.performHkOperation() == returnvalue::OK);
|
||||
/* Now three messages should be sent. */
|
||||
REQUIRE(poolOwner.hkHelper.performHkOperation() == returnvalue::OK);
|
||||
// Now three messages should be sent.
|
||||
REQUIRE(poolOwnerMock.wasMessageSent());
|
||||
CHECK(poolOwnerMock.numberOfSentMessages() == 3);
|
||||
CHECK(poolOwnerMock.numberOfSentMessagesToDest(subscriberId) == 2);
|
||||
@ -267,20 +267,20 @@ TEST_CASE("Local Pool Manager Tests", "[LocManTest]") {
|
||||
}
|
||||
|
||||
SECTION("PeriodicHKAndMessaging") {
|
||||
/* Now we subcribe for a HK periodic generation. Even when it's difficult to simulate
|
||||
the temporal behaviour correctly the HK manager should generate a HK packet
|
||||
immediately and the periodic helper depends on HK op function calls anyway instead of
|
||||
using the clock, so we could also just call performHkOperation multiple times */
|
||||
//Now we subcribe for a HK periodic generation. Even when it's difficult to simulate
|
||||
//the temporal behaviour correctly the HK manager should generate a HK packet
|
||||
//immediately and the periodic helper depends on HK op function calls anyway instead of
|
||||
//using the clock, so we could also just call performHkOperation multiple times
|
||||
REQUIRE(poolOwner.subscribePeriodicHk(true) == returnvalue::OK);
|
||||
REQUIRE(poolOwner.poolManager.performHkOperation() == returnvalue::OK);
|
||||
/* Now HK packet should be sent as message immediately. */
|
||||
REQUIRE(poolOwner.hkHelper.performHkOperation() == returnvalue::OK);
|
||||
// Now HK packet should be sent as message immediately.
|
||||
REQUIRE(poolOwnerMock.wasMessageSent());
|
||||
CHECK(poolOwnerMock.numberOfSentMessages() == 1);
|
||||
CHECK(poolOwnerMock.clearLastSentMessage() == returnvalue::OK);
|
||||
|
||||
LocalPoolDataSetBase* setHandle = poolOwner.getDataSetHandle(lpool::testSid);
|
||||
REQUIRE(setHandle != nullptr);
|
||||
CHECK(poolOwner.poolManager.generateHousekeepingPacket(lpool::testSid, setHandle, false) ==
|
||||
CHECK(poolOwner.hkHelper.generateHousekeepingPacket(lpool::testSid, setHandle, false) ==
|
||||
returnvalue::OK);
|
||||
REQUIRE(poolOwnerMock.wasMessageSent());
|
||||
CHECK(poolOwnerMock.numberOfSentMessages() == 1);
|
||||
@ -289,41 +289,41 @@ TEST_CASE("Local Pool Manager Tests", "[LocManTest]") {
|
||||
CHECK(setHandle->getReportingEnabled() == true);
|
||||
CommandMessage hkCmd;
|
||||
HousekeepingMessage::setToggleReportingCommand(&hkCmd, lpool::testSid, false);
|
||||
CHECK(poolOwner.poolManager.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(poolOwner.hkHelper.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(setHandle->getReportingEnabled() == false);
|
||||
REQUIRE(poolOwnerMock.wasMessageSent());
|
||||
CHECK(poolOwnerMock.numberOfSentMessages() == 1);
|
||||
CHECK(poolOwnerMock.clearLastSentMessage() == returnvalue::OK);
|
||||
|
||||
HousekeepingMessage::setToggleReportingCommand(&hkCmd, lpool::testSid, true);
|
||||
CHECK(poolOwner.poolManager.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(poolOwner.hkHelper.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(setHandle->getReportingEnabled() == true);
|
||||
REQUIRE(poolOwnerMock.wasMessageSent());
|
||||
CHECK(poolOwnerMock.clearLastSentMessage() == returnvalue::OK);
|
||||
|
||||
HousekeepingMessage::setToggleReportingCommand(&hkCmd, lpool::testSid, false);
|
||||
CHECK(poolOwner.poolManager.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(poolOwner.hkHelper.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(setHandle->getReportingEnabled() == false);
|
||||
REQUIRE(poolOwnerMock.wasMessageSent());
|
||||
CHECK(poolOwnerMock.clearLastSentMessage() == returnvalue::OK);
|
||||
|
||||
HousekeepingMessage::setCollectionIntervalModificationCommand(&hkCmd, lpool::testSid, 0.4);
|
||||
CHECK(poolOwner.poolManager.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(poolOwner.hkHelper.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK_THAT(poolOwner.dataset.getCollectionInterval(), Catch::Matchers::WithinAbs(0.4, 0.01));
|
||||
REQUIRE(poolOwnerMock.wasMessageSent());
|
||||
REQUIRE(poolOwnerMock.numberOfSentMessages() == 1);
|
||||
CHECK(poolOwnerMock.clearLastSentMessage() == returnvalue::OK);
|
||||
|
||||
HousekeepingMessage::setStructureReportingCommand(&hkCmd, lpool::testSid);
|
||||
REQUIRE(poolOwner.poolManager.performHkOperation() == returnvalue::OK);
|
||||
CHECK(poolOwner.poolManager.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
/* Now HK packet should be sent as message. */
|
||||
REQUIRE(poolOwner.hkHelper.performHkOperation() == returnvalue::OK);
|
||||
CHECK(poolOwner.hkHelper.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
// Now HK packet should be sent as message.
|
||||
REQUIRE(poolOwnerMock.wasMessageSent());
|
||||
REQUIRE(poolOwnerMock.numberOfSentMessages() == 1);
|
||||
poolOwnerMock.clearMessages();
|
||||
|
||||
HousekeepingMessage::setOneShotReportCommand(&hkCmd, lpool::testSid);
|
||||
CHECK(poolOwner.poolManager.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(poolOwner.hkHelper.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
REQUIRE(poolOwnerMock.wasMessageSent());
|
||||
REQUIRE(poolOwnerMock.numberOfSentMessages() == 1);
|
||||
poolOwnerMock.clearMessages();
|
||||
@ -331,63 +331,64 @@ TEST_CASE("Local Pool Manager Tests", "[LocManTest]") {
|
||||
HousekeepingMessage::setUpdateNotificationSetCommand(&hkCmd, lpool::testSid);
|
||||
sid_t sidToCheck;
|
||||
store_address_t storeId;
|
||||
CHECK(poolOwner.poolManager.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(poolOwner.hkHelper.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(poolOwner.changedDataSetCallbackWasCalled(sidToCheck, storeId) == true);
|
||||
CHECK(sidToCheck == lpool::testSid);
|
||||
|
||||
HousekeepingMessage::setStructureReportingCommand(&hkCmd, lpool::testSid);
|
||||
CHECK(poolOwner.poolManager.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(poolOwner.hkHelper.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
REQUIRE(poolOwnerMock.wasMessageSent());
|
||||
REQUIRE(poolOwnerMock.numberOfSentMessages() == 1);
|
||||
poolOwnerMock.clearMessages();
|
||||
|
||||
HousekeepingMessage::setCollectionIntervalModificationCommand(&hkCmd, lpool::testSid, 0.4);
|
||||
CHECK(poolOwner.poolManager.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(poolOwner.hkHelper.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
REQUIRE(poolOwnerMock.wasMessageSent());
|
||||
REQUIRE(poolOwnerMock.numberOfSentMessages() == 1);
|
||||
poolOwnerMock.clearMessages();
|
||||
|
||||
HousekeepingMessage::setToggleReportingCommand(&hkCmd, lpool::testSid, true);
|
||||
CHECK(poolOwner.poolManager.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(poolOwner.hkHelper.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
REQUIRE(poolOwnerMock.wasMessageSent());
|
||||
REQUIRE(poolOwnerMock.numberOfSentMessages() == 1);
|
||||
poolOwnerMock.clearMessages();
|
||||
|
||||
HousekeepingMessage::setToggleReportingCommand(&hkCmd, lpool::testSid, false);
|
||||
CHECK(poolOwner.poolManager.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(poolOwner.hkHelper.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
REQUIRE(poolOwnerMock.wasMessageSent());
|
||||
REQUIRE(poolOwnerMock.numberOfSentMessages() == 1);
|
||||
poolOwnerMock.clearMessages();
|
||||
|
||||
HousekeepingMessage::setOneShotReportCommand(&hkCmd, lpool::testSid);
|
||||
CHECK(poolOwner.poolManager.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(poolOwner.hkHelper.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
REQUIRE(poolOwnerMock.wasMessageSent());
|
||||
REQUIRE(poolOwnerMock.numberOfSentMessages() == 1);
|
||||
poolOwnerMock.clearMessages();
|
||||
|
||||
HousekeepingMessage::setUpdateNotificationVariableCommand(&hkCmd, lpool::uint8VarGpid);
|
||||
gp_id_t gpidToCheck;
|
||||
CHECK(poolOwner.poolManager.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(poolOwner.hkHelper.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(poolOwner.changedVariableCallbackWasCalled(gpidToCheck, storeId) == true);
|
||||
CHECK(gpidToCheck == lpool::uint8VarGpid);
|
||||
|
||||
HousekeepingMessage::setUpdateSnapshotSetCommand(&hkCmd, lpool::testSid,
|
||||
store_address_t::invalid());
|
||||
CHECK(poolOwner.poolManager.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(poolOwner.hkHelper.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(poolOwner.changedDataSetCallbackWasCalled(sidToCheck, storeId) == true);
|
||||
CHECK(sidToCheck == lpool::testSid);
|
||||
|
||||
HousekeepingMessage::setUpdateSnapshotVariableCommand(&hkCmd, lpool::uint8VarGpid,
|
||||
store_address_t::invalid());
|
||||
CHECK(poolOwner.poolManager.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(poolOwner.hkHelper.handleHousekeepingMessage(&hkCmd) == returnvalue::OK);
|
||||
CHECK(poolOwner.changedVariableCallbackWasCalled(gpidToCheck, storeId) == true);
|
||||
CHECK(gpidToCheck == lpool::uint8VarGpid);
|
||||
|
||||
poolOwner.poolManager.printPoolEntry(lpool::uint8VarId);
|
||||
poolOwner.hkHelper.printPoolEntry(lpool::uint8VarId);
|
||||
}
|
||||
|
||||
/* we need to reset the subscription list because the pool owner
|
||||
is a global object. */
|
||||
// we need to reset the subscription list because the pool owner
|
||||
//is a global object.
|
||||
CHECK(poolOwner.reset() == returnvalue::OK);
|
||||
poolOwnerMock.clearMessages(true);
|
||||
*/
|
||||
}
|
||||
|
@ -1,4 +1,4 @@
|
||||
#include <fsfw/datapoollocal/HasLocalDataPoolIF.h>
|
||||
#include <fsfw/datapoollocal/PeriodicHkGenerationHelper.h>
|
||||
#include <fsfw/objectmanager/ObjectManager.h>
|
||||
|
||||
#include <catch2/catch_test_macros.hpp>
|
||||
@ -12,8 +12,7 @@ using namespace returnvalue;
|
||||
TEST_CASE("LocalPoolVariable", "[LocPoolVarTest]") {
|
||||
auto queue = MessageQueueMock(1);
|
||||
LocalPoolOwnerBase poolOwner(queue, objects::TEST_LOCAL_POOL_OWNER_BASE);
|
||||
REQUIRE(poolOwner.initializeHkManager() == OK);
|
||||
REQUIRE(poolOwner.initializeHkManagerAfterTaskCreation() == OK);
|
||||
REQUIRE(poolOwner.initialize() == OK);
|
||||
|
||||
SECTION("Basic Tests") {
|
||||
/* very basic test. */
|
||||
@ -25,10 +24,7 @@ TEST_CASE("LocalPoolVariable", "[LocPoolVarTest]") {
|
||||
REQUIRE(testVariable.commit() == returnvalue::OK);
|
||||
REQUIRE(testVariable.read() == returnvalue::OK);
|
||||
REQUIRE(testVariable.value == 5);
|
||||
CHECK(not testVariable.isValid());
|
||||
testVariable.setValid(true);
|
||||
CHECK(testVariable.isValid());
|
||||
CHECK(testVariable.commit(true) == returnvalue::OK);
|
||||
CHECK(testVariable.commit() == returnvalue::OK);
|
||||
|
||||
testVariable.setReadWriteMode(pool_rwm_t::VAR_READ);
|
||||
CHECK(testVariable.getReadWriteMode() == pool_rwm_t::VAR_READ);
|
||||
@ -38,10 +34,6 @@ TEST_CASE("LocalPoolVariable", "[LocPoolVarTest]") {
|
||||
CHECK(testVariable.getDataPoolId() == 22);
|
||||
testVariable.setDataPoolId(lpool::uint8VarId);
|
||||
|
||||
testVariable.setChanged(true);
|
||||
CHECK(testVariable.hasChanged());
|
||||
testVariable.setChanged(false);
|
||||
|
||||
gp_id_t globPoolId(objects::TEST_LOCAL_POOL_OWNER_BASE, lpool::uint8VarId);
|
||||
lp_var_t<uint8_t> testVariable2 = lp_var_t<uint8_t>(globPoolId);
|
||||
REQUIRE(testVariable2.read() == returnvalue::OK);
|
||||
@ -105,6 +97,5 @@ TEST_CASE("LocalPoolVariable", "[LocPoolVarTest]") {
|
||||
lp_var_t<uint8_t> invalidObjectVar = lp_var_t<uint8_t>(0xffffffff, lpool::uint8VarId);
|
||||
gp_id_t globPoolId(0xffffffff, lpool::uint8VarId);
|
||||
lp_var_t<uint8_t> invalidObjectVar2 = lp_var_t<uint8_t>(globPoolId);
|
||||
lp_var_t<uint8_t> invalidObjectVar3 = lp_var_t<uint8_t>(nullptr, lpool::uint8VarId);
|
||||
}
|
||||
}
|
||||
|
@ -1,4 +1,4 @@
|
||||
#include <fsfw/datapoollocal/HasLocalDataPoolIF.h>
|
||||
#include <fsfw/datapoollocal/PeriodicHkGenerationHelper.h>
|
||||
#include <fsfw/objectmanager/ObjectManager.h>
|
||||
|
||||
#include <catch2/catch_test_macros.hpp>
|
||||
@ -12,8 +12,7 @@ using namespace returnvalue;
|
||||
TEST_CASE("LocalPoolVector", "[LocPoolVecTest]") {
|
||||
auto queue = MessageQueueMock(1);
|
||||
LocalPoolOwnerBase poolOwner(queue, objects::TEST_LOCAL_POOL_OWNER_BASE);
|
||||
REQUIRE(poolOwner.initializeHkManager() == OK);
|
||||
REQUIRE(poolOwner.initializeHkManagerAfterTaskCreation() == OK);
|
||||
REQUIRE(poolOwner.initialize() == OK);
|
||||
|
||||
SECTION("BasicTest") {
|
||||
// very basic test.
|
||||
@ -25,7 +24,7 @@ TEST_CASE("LocalPoolVector", "[LocPoolVecTest]") {
|
||||
testVector.value[2] = 32023;
|
||||
|
||||
REQUIRE(testVector.commit(true) == returnvalue::OK);
|
||||
CHECK(testVector.isValid());
|
||||
// CHECK(testVector.isValid());
|
||||
|
||||
testVector.value[0] = 0;
|
||||
testVector.value[1] = 0;
|
||||
|
@ -31,8 +31,8 @@ TEST_CASE("Internal Error Reporter", "[TestInternalError]") {
|
||||
task.startTask();
|
||||
MessageQueueIF* testQueue = QueueFactory::instance()->createMessageQueue(1);
|
||||
MessageQueueIF* hkQueue = QueueFactory::instance()->createMessageQueue(1);
|
||||
internalErrorReporter->getSubscriptionInterface()->subscribeForSetUpdateMessage(
|
||||
InternalErrorDataset::ERROR_SET_ID, objects::NO_OBJECT, hkQueue->getId(), true);
|
||||
// internalErrorReporter->getSubscriptionInterface()->subscribeForSetUpdateMessage(
|
||||
// InternalErrorDataset::ERROR_SET_ID, objects::NO_OBJECT, hkQueue->getId(), true);
|
||||
auto* ipcStore = ObjectManager::instance()->get<StorageManagerIF>(objects::IPC_STORE);
|
||||
SECTION("MessageQueueFull") {
|
||||
CommandMessage message;
|
||||
@ -46,6 +46,8 @@ TEST_CASE("Internal Error Reporter", "[TestInternalError]") {
|
||||
then remeber the queueHit count and force another hit */
|
||||
internalErrorReporter->queueMessageNotSent();
|
||||
internalErrorReporter->performOperation(0);
|
||||
// TODO: Fix test
|
||||
/*
|
||||
{
|
||||
CommandMessage hkMessage;
|
||||
result = hkQueue->receiveMessage(&hkMessage);
|
||||
@ -114,6 +116,7 @@ TEST_CASE("Internal Error Reporter", "[TestInternalError]") {
|
||||
REQUIRE(result == returnvalue::OK);
|
||||
internalErrorReporter->performOperation(0);
|
||||
}
|
||||
*/
|
||||
}
|
||||
QueueFactory::instance()->deleteMessageQueue(testQueue);
|
||||
QueueFactory::instance()->deleteMessageQueue(hkQueue);
|
||||
|
@ -108,3 +108,14 @@ ReturnValue_t DeviceHandlerMock::initialize() {
|
||||
setMode(MODE_ON);
|
||||
return result;
|
||||
}
|
||||
|
||||
ReturnValue_t DeviceHandlerMock::serializeDataset(sid_t structureId, uint8_t *buf, size_t maxSize) {
|
||||
return returnvalue::OK;
|
||||
}
|
||||
|
||||
ReturnValue_t DeviceHandlerMock::specifyDatasets(
|
||||
std::vector<periodicHk::SetSpecification> &setList) {
|
||||
return returnvalue::OK;
|
||||
}
|
||||
|
||||
localpool::SharedPool *DeviceHandlerMock::getOptionalSharedPool() { return nullptr; }
|
||||
|
@ -23,6 +23,12 @@ class DeviceHandlerMock : public DeviceHandlerBase {
|
||||
|
||||
ReturnValue_t initialize() override;
|
||||
|
||||
ReturnValue_t serializeDataset(sid_t structureId, uint8_t *buf, size_t maxSize) override;
|
||||
|
||||
ReturnValue_t specifyDatasets(std::vector<periodicHk::SetSpecification> &setList) override;
|
||||
|
||||
localpool::SharedPool *getOptionalSharedPool() override;
|
||||
|
||||
protected:
|
||||
void doStartUp() override;
|
||||
void doShutDown() override;
|
||||
|
@ -3,19 +3,38 @@
|
||||
LocalPoolOwnerBase::LocalPoolOwnerBase(MessageQueueIF &queue, object_id_t objectId)
|
||||
: SystemObject(objectId),
|
||||
queue(queue),
|
||||
poolManager(this, &queue),
|
||||
dataset(this, lpool::testSetId) {}
|
||||
sharedPool(getObjectId()),
|
||||
hkHelper(this, &queue),
|
||||
dataset(sharedPool, lpool::testSetId) {}
|
||||
|
||||
LocalPoolOwnerBase::~LocalPoolOwnerBase() = default;
|
||||
|
||||
ReturnValue_t LocalPoolOwnerBase::initializeHkManager() {
|
||||
if (not initialized) {
|
||||
initialized = true;
|
||||
return poolManager.initialize(&queue);
|
||||
ReturnValue_t LocalPoolOwnerBase::initialize() {
|
||||
sharedPool.addPoolEntry(lpool::uint8VarId, &u8PoolEntry);
|
||||
sharedPool.addPoolEntry(lpool::floatVarId, &floatPoolEntry);
|
||||
sharedPool.addPoolEntry(lpool::uint32VarId, &u32PoolEntry);
|
||||
sharedPool.addPoolEntry(lpool::uint16Vec3Id, &u16VecPoolEntry);
|
||||
sharedPool.addPoolEntry(lpool::int64Vec2Id, &i64VecPoolEntry);
|
||||
ReturnValue_t result = hkHelper.initialize(&queue);
|
||||
if (result != returnvalue::OK) {
|
||||
return result;
|
||||
}
|
||||
return SystemObject::initialize();
|
||||
}
|
||||
|
||||
localpool::SharedPool *LocalPoolOwnerBase::getOptionalSharedPool() { return &sharedPool; }
|
||||
|
||||
ReturnValue_t LocalPoolOwnerBase::serializeDataset(sid_t structureId, uint8_t *buf,
|
||||
size_t maxSize) {
|
||||
return returnvalue::OK;
|
||||
}
|
||||
|
||||
ReturnValue_t LocalPoolOwnerBase::specifyDatasets(
|
||||
std::vector<periodicHk::SetSpecification> &setList) {
|
||||
return returnvalue::OK;
|
||||
}
|
||||
|
||||
/*
|
||||
ReturnValue_t LocalPoolOwnerBase::initializeLocalDataPool(localpool::DataPool &localDataPoolMap,
|
||||
PeriodicHkGenerationHelper &poolManager) {
|
||||
// Default initialization empty for now.
|
||||
@ -27,6 +46,7 @@ ReturnValue_t LocalPoolOwnerBase::initializeLocalDataPool(localpool::DataPool &l
|
||||
localDataPoolMap.emplace(lpool::int64Vec2Id, &i64VecPoolEntry);
|
||||
return returnvalue::OK;
|
||||
}
|
||||
*/
|
||||
|
||||
LocalPoolObjectBase *LocalPoolOwnerBase::getPoolObjectHandle(lp_id_t localPoolId) {
|
||||
if (localPoolId == lpool::uint8VarId) {
|
||||
@ -93,14 +113,6 @@ bool LocalPoolOwnerBase::changedDataSetCallbackWasCalled(sid_t &sid, store_addre
|
||||
return condition;
|
||||
}
|
||||
|
||||
/*
|
||||
void LocalPoolOwnerBase::handleChangedDataset(sid_t sid, store_address_t storeId,
|
||||
bool *clearMessage) {
|
||||
this->changedDatasetSid = sid;
|
||||
this->storeIdForChangedSet = storeId;
|
||||
}
|
||||
*/
|
||||
|
||||
bool LocalPoolOwnerBase::changedVariableCallbackWasCalled(gp_id_t &gpid, store_address_t &storeId) {
|
||||
bool condition = false;
|
||||
if (not this->changedPoolVariableGpid.notSet()) {
|
||||
@ -113,20 +125,4 @@ bool LocalPoolOwnerBase::changedVariableCallbackWasCalled(gp_id_t &gpid, store_a
|
||||
return condition;
|
||||
}
|
||||
|
||||
ReturnValue_t LocalPoolOwnerBase::initializeHkManagerAfterTaskCreation() {
|
||||
if (not initializedAfterTaskCreation) {
|
||||
initializedAfterTaskCreation = true;
|
||||
return poolManager.initializeAfterTaskCreation();
|
||||
}
|
||||
return returnvalue::OK;
|
||||
}
|
||||
|
||||
/*
|
||||
void LocalPoolOwnerBase::handleChangedPoolVariable(gp_id_t globPoolId, store_address_t storeId,
|
||||
bool *clearMessage) {
|
||||
this->changedPoolVariableGpid = globPoolId;
|
||||
this->storeIdForChangedVariable = storeId;
|
||||
}
|
||||
*/
|
||||
|
||||
void LocalPoolOwnerBase::setHkDestId(MessageQueueId_t id) { poolManager.setHkDestinationId(id); }
|
||||
void LocalPoolOwnerBase::setHkDestId(MessageQueueId_t id) { hkHelper.setHkDestinationId(id); }
|
||||
|
@ -2,10 +2,10 @@
|
||||
#define FSFW_UNITTEST_TESTS_DATAPOOLLOCAL_LOCALPOOLOWNERBASE_H_
|
||||
|
||||
#include <fsfw/datapool/PoolReadGuard.h>
|
||||
#include <fsfw/datapoollocal/HasLocalDataPoolIF.h>
|
||||
#include <fsfw/datapoollocal/LocalDataSet.h>
|
||||
#include <fsfw/datapoollocal/LocalPoolVariable.h>
|
||||
#include <fsfw/datapoollocal/LocalPoolVector.h>
|
||||
#include <fsfw/datapoollocal/PeriodicHkGenerationHelper.h>
|
||||
#include <fsfw/datapoollocal/StaticLocalDataSet.h>
|
||||
#include <fsfw/ipc/QueueFactory.h>
|
||||
#include <fsfw/objectmanager/SystemObject.h>
|
||||
@ -37,8 +37,8 @@ class LocalPoolStaticTestDataSet : public StaticLocalDataSet<3> {
|
||||
public:
|
||||
LocalPoolStaticTestDataSet() : StaticLocalDataSet(lpool::testSid) {}
|
||||
|
||||
LocalPoolStaticTestDataSet(PeriodicHkGenerationIF* owner, uint32_t setId)
|
||||
: StaticLocalDataSet(owner, setId) {}
|
||||
LocalPoolStaticTestDataSet(localpool::SharedPool& sharedPool, uint32_t setId)
|
||||
: StaticLocalDataSet(sharedPool, setId) {}
|
||||
|
||||
lp_var_t<uint8_t> localPoolVarUint8 = lp_var_t<uint8_t>(lpool::uint8VarGpid, this);
|
||||
lp_var_t<float> localPoolVarFloat = lp_var_t<float>(lpool::floatVarGpid, this);
|
||||
@ -51,8 +51,8 @@ class LocalPoolTestDataSet : public LocalDataSet {
|
||||
public:
|
||||
LocalPoolTestDataSet() : LocalDataSet(lpool::testSid, lpool::dataSetMaxVariables) {}
|
||||
|
||||
LocalPoolTestDataSet(PeriodicHkGenerationIF* owner, uint32_t setId)
|
||||
: LocalDataSet(owner, setId, lpool::dataSetMaxVariables) {}
|
||||
LocalPoolTestDataSet(localpool::SharedPool& sharedPool, uint32_t setId)
|
||||
: LocalDataSet(sharedPool, setId, lpool::dataSetMaxVariables) {}
|
||||
|
||||
lp_var_t<uint8_t> localPoolVarUint8 = lp_var_t<uint8_t>(lpool::uint8VarGpid, this);
|
||||
lp_var_t<float> localPoolVarFloat = lp_var_t<float>(lpool::floatVarGpid, this);
|
||||
@ -70,31 +70,20 @@ class LocalPoolOwnerBase : public SystemObject, public PeriodicHkGenerationIF {
|
||||
|
||||
[[nodiscard]] object_id_t getObjectId() const override { return SystemObject::getObjectId(); }
|
||||
|
||||
ReturnValue_t initializeHkManager();
|
||||
ReturnValue_t serializeDataset(sid_t structureId, uint8_t* buf, size_t maxSize) override;
|
||||
|
||||
ReturnValue_t specifyDatasets(std::vector<periodicHk::SetSpecification>& setList) override;
|
||||
|
||||
localpool::SharedPool* getOptionalSharedPool() override;
|
||||
ReturnValue_t initialize() override;
|
||||
|
||||
void setHkDestId(MessageQueueId_t id);
|
||||
|
||||
ReturnValue_t initializeHkManagerAfterTaskCreation();
|
||||
// ReturnValue_t initializeHkManagerAfterTaskCreation();
|
||||
|
||||
/** Command queue for housekeeping messages. */
|
||||
[[nodiscard]] MessageQueueId_t getCommandQueue() const override { return queue.getId(); }
|
||||
|
||||
// This is called by initializeAfterTaskCreation of the HK manager.
|
||||
ReturnValue_t initializeLocalDataPool(localpool::DataPool& localDataPoolMap,
|
||||
PeriodicHkGenerationHelper& poolManager) override;
|
||||
|
||||
PeriodicHkGenerationHelper* getHkManagerHandle() override { return &poolManager; }
|
||||
|
||||
//[[nodiscard]] dur_millis_t getPeriodicOperationFrequency() const override { return 200; }
|
||||
|
||||
/**
|
||||
* This function is used by the pool manager to get a valid dataset
|
||||
* from a SID
|
||||
* @param sid Corresponding structure ID
|
||||
* @return
|
||||
*/
|
||||
LocalPoolDataSetBase* getDataSetHandle(sid_t sid) override { return &dataset; }
|
||||
|
||||
LocalPoolObjectBase* getPoolObjectHandle(lp_id_t localPoolId) override;
|
||||
|
||||
[[nodiscard]] MessageQueueMock& getMockQueueHandle() const {
|
||||
@ -102,64 +91,22 @@ class LocalPoolOwnerBase : public SystemObject, public PeriodicHkGenerationIF {
|
||||
}
|
||||
|
||||
ReturnValue_t subscribePeriodicHk(bool enableReporting) {
|
||||
return poolManager.enableRegularPeriodicPacket(
|
||||
subdp::RegularHkPeriodicParams(lpool::testSid, enableReporting, 0.2));
|
||||
}
|
||||
|
||||
ReturnValue_t subscribeWrapperSetUpdate(MessageQueueId_t receiverId) {
|
||||
return poolManager.subscribeForSetUpdateMessage(lpool::testSetId, objects::NO_OBJECT,
|
||||
receiverId, false);
|
||||
}
|
||||
|
||||
ReturnValue_t subscribeWrapperSetUpdateSnapshot(MessageQueueId_t receiverId) {
|
||||
return poolManager.subscribeForSetUpdateMessage(lpool::testSetId, objects::NO_OBJECT,
|
||||
receiverId, true);
|
||||
}
|
||||
|
||||
ReturnValue_t subscribeWrapperSetUpdateHk(bool diagnostics = false,
|
||||
AcceptsHkPacketsIF* receiver = nullptr) {
|
||||
if (diagnostics) {
|
||||
auto params = subdp::DiagnosticsHkUpdateParams(lpool::testSid, true);
|
||||
if (receiver != nullptr) {
|
||||
params.receiver = receiver->getHkQueue();
|
||||
}
|
||||
return poolManager.subscribeForDiagUpdatePacket(params);
|
||||
} else {
|
||||
auto params = subdp::RegularHkUpdateParams(lpool::testSid, true);
|
||||
if (receiver != nullptr) {
|
||||
params.receiver = receiver->getHkQueue();
|
||||
}
|
||||
return poolManager.subscribeForRegularUpdatePacket(params);
|
||||
}
|
||||
}
|
||||
|
||||
ReturnValue_t subscribeWrapperVariableUpdate(MessageQueueId_t receiverId, lp_id_t localPoolId) {
|
||||
return poolManager.subscribeForVariableUpdateMessage(localPoolId, MessageQueueIF::NO_QUEUE,
|
||||
receiverId, false);
|
||||
}
|
||||
|
||||
ReturnValue_t subscribeWrapperVariableSnapshot(MessageQueueId_t receiverId, lp_id_t localPoolId) {
|
||||
return poolManager.subscribeForVariableUpdateMessage(localPoolId, MessageQueueIF::NO_QUEUE,
|
||||
receiverId, true);
|
||||
return hkHelper.enablePeriodicPacket(lpool::testSid, 200);
|
||||
}
|
||||
|
||||
ReturnValue_t reset();
|
||||
|
||||
// void resetSubscriptionList() { poolManager.clearReceiversList(); }
|
||||
|
||||
bool changedDataSetCallbackWasCalled(sid_t& sid, store_address_t& storeId);
|
||||
bool changedVariableCallbackWasCalled(gp_id_t& gpid, store_address_t& storeId);
|
||||
|
||||
PeriodicHkGenerationHelper poolManager;
|
||||
PeriodicHkGenerationHelper hkHelper;
|
||||
localpool::SharedPool sharedPool;
|
||||
LocalPoolTestDataSet dataset;
|
||||
|
||||
private:
|
||||
// void handleChangedDataset(sid_t sid, store_address_t storeId, bool* clearMessage) override;
|
||||
sid_t changedDatasetSid;
|
||||
store_address_t storeIdForChangedSet;
|
||||
|
||||
// void handleChangedPoolVariable(gp_id_t globPoolId, store_address_t storeId,
|
||||
// bool* clearMessage) override;
|
||||
gp_id_t changedPoolVariableGpid;
|
||||
store_address_t storeIdForChangedVariable;
|
||||
|
||||
@ -169,11 +116,11 @@ class LocalPoolOwnerBase : public SystemObject, public PeriodicHkGenerationIF {
|
||||
PoolEntry<uint16_t> u16VecPoolEntry = PoolEntry<uint16_t>({0, 0, 0});
|
||||
PoolEntry<int64_t> i64VecPoolEntry = PoolEntry<int64_t>({0, 0});
|
||||
|
||||
lp_var_t<uint8_t> testUint8 = lp_var_t<uint8_t>(this, lpool::uint8VarId);
|
||||
lp_var_t<float> testFloat = lp_var_t<float>(this, lpool::floatVarId);
|
||||
lp_var_t<uint32_t> testUint32 = lp_var_t<uint32_t>(this, lpool::uint32VarId);
|
||||
lp_vec_t<uint16_t, 3> testUint16Vec = lp_vec_t<uint16_t, 3>(this, lpool::uint16Vec3Id);
|
||||
lp_vec_t<int64_t, 2> testInt64Vec = lp_vec_t<int64_t, 2>(this, lpool::int64Vec2Id);
|
||||
lp_var_t<uint8_t> testUint8 = lp_var_t<uint8_t>(sharedPool, lpool::uint8VarId);
|
||||
lp_var_t<float> testFloat = lp_var_t<float>(sharedPool, lpool::floatVarId);
|
||||
lp_var_t<uint32_t> testUint32 = lp_var_t<uint32_t>(sharedPool, lpool::uint32VarId);
|
||||
lp_vec_t<uint16_t, 3> testUint16Vec = lp_vec_t<uint16_t, 3>(sharedPool, lpool::uint16Vec3Id);
|
||||
lp_vec_t<int64_t, 2> testInt64Vec = lp_vec_t<int64_t, 2>(sharedPool, lpool::int64Vec2Id);
|
||||
|
||||
MessageQueueIF& queue;
|
||||
|
||||
|
Reference in New Issue
Block a user