Mercurial > hg > orthanc
view OrthancServer/ServerIndex.cpp @ 597:11a195070841 Orthanc-0.6.2
set version
author | Sebastien Jodogne <s.jodogne@gmail.com> |
---|---|
date | Fri, 04 Oct 2013 16:33:53 +0200 |
parents | b2357f1f026f |
children | ec0b7a51d7bd |
line wrap: on
line source
/** * Orthanc - A Lightweight, RESTful DICOM Store * Copyright (C) 2012-2013 Medical Physics Department, CHU of Liege, * Belgium * * This program is free software: you can redistribute it and/or * modify it under the terms of the GNU General Public License as * published by the Free Software Foundation, either version 3 of the * License, or (at your option) any later version. * * In addition, as a special exception, the copyright holders of this * program give permission to link the code of its release with the * OpenSSL project's "OpenSSL" library (or with modified versions of it * that use the same license as the "OpenSSL" library), and distribute * the linked executables. You must obey the GNU General Public License * in all respects for all of the code used other than "OpenSSL". If you * modify file(s) with this exception, you may extend this exception to * your version of the file(s), but you are not obligated to do so. If * you do not wish to do so, delete this exception statement from your * version. If you delete this exception statement from all source files * in the program, then also delete it here. * * This program is distributed in the hope that it will be useful, but * WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program. If not, see <http://www.gnu.org/licenses/>. **/ #include "ServerIndex.h" #ifndef NOMINMAX #define NOMINMAX #endif #include "EmbeddedResources.h" #include "OrthancInitialization.h" #include "../Core/Toolbox.h" #include "../Core/Uuid.h" #include "../Core/DicomFormat/DicomArray.h" #include "../Core/SQLite/Transaction.h" #include "FromDcmtkBridge.h" #include "ServerContext.h" #include <boost/lexical_cast.hpp> #include <stdio.h> #include <glog/logging.h> static const uint64_t MEGA_BYTES = 1024 * 1024; namespace Orthanc { namespace Internals { class ServerIndexListener : public IServerIndexListener { private: ServerContext& context_; bool hasRemainingLevel_; ResourceType remainingType_; std::string remainingPublicId_; std::list<std::string> pendingFilesToRemove_; uint64_t sizeOfFilesToRemove_; public: ServerIndexListener(ServerContext& context) : context_(context) { Reset(); assert(ResourceType_Patient < ResourceType_Study && ResourceType_Study < ResourceType_Series && ResourceType_Series < ResourceType_Instance); } void Reset() { sizeOfFilesToRemove_ = 0; hasRemainingLevel_ = false; pendingFilesToRemove_.clear(); } uint64_t GetSizeOfFilesToRemove() { return sizeOfFilesToRemove_; } void CommitFilesToRemove() { for (std::list<std::string>::iterator it = pendingFilesToRemove_.begin(); it != pendingFilesToRemove_.end(); it++) { context_.RemoveFile(*it); } } virtual void SignalRemainingAncestor(ResourceType parentType, const std::string& publicId) { LOG(INFO) << "Remaining ancestor \"" << publicId << "\" (" << parentType << ")"; if (hasRemainingLevel_) { if (parentType < remainingType_) { remainingType_ = parentType; remainingPublicId_ = publicId; } } else { hasRemainingLevel_ = true; remainingType_ = parentType; remainingPublicId_ = publicId; } } virtual void SignalFileDeleted(const FileInfo& info) { assert(Toolbox::IsUuid(info.GetUuid())); pendingFilesToRemove_.push_back(info.GetUuid()); sizeOfFilesToRemove_ += info.GetCompressedSize(); } bool HasRemainingLevel() const { return hasRemainingLevel_; } ResourceType GetRemainingType() const { assert(HasRemainingLevel()); return remainingType_; } const std::string& GetRemainingPublicId() const { assert(HasRemainingLevel()); return remainingPublicId_; } }; } class ServerIndex::Transaction { private: ServerIndex& index_; std::auto_ptr<SQLite::Transaction> transaction_; bool isCommitted_; public: Transaction(ServerIndex& index) : index_(index), isCommitted_(false) { assert(index_.currentStorageSize_ == index_.db_->GetTotalCompressedSize()); index_.listener_->Reset(); transaction_.reset(index_.db_->StartTransaction()); transaction_->Begin(); } void Commit(uint64_t sizeOfAddedFiles) { if (!isCommitted_) { transaction_->Commit(); // We can remove the files once the SQLite transaction has // been successfully committed. Some files might have to be // deleted because of recycling. index_.listener_->CommitFilesToRemove(); index_.currentStorageSize_ += sizeOfAddedFiles; assert(index_.currentStorageSize_ >= index_.listener_->GetSizeOfFilesToRemove()); index_.currentStorageSize_ -= index_.listener_->GetSizeOfFilesToRemove(); assert(index_.currentStorageSize_ == index_.db_->GetTotalCompressedSize()); isCommitted_ = true; } } }; struct ServerIndex::UnstableResourcePayload { Orthanc::ResourceType type_; boost::posix_time::ptime time_; UnstableResourcePayload() : type_(Orthanc::ResourceType_Instance) { } UnstableResourcePayload(Orthanc::ResourceType type) : type_(type) { time_ = boost::posix_time::second_clock::local_time(); } unsigned int GetAge() const { return (boost::posix_time::second_clock::local_time() - time_).total_seconds(); } }; bool ServerIndex::DeleteResource(Json::Value& target, const std::string& uuid, ResourceType expectedType) { boost::mutex::scoped_lock lock(mutex_); listener_->Reset(); Transaction t(*this); int64_t id; ResourceType type; if (!db_->LookupResource(uuid, id, type) || expectedType != type) { return false; } db_->DeleteResource(id); if (listener_->HasRemainingLevel()) { ResourceType type = listener_->GetRemainingType(); const std::string& uuid = listener_->GetRemainingPublicId(); target["RemainingAncestor"] = Json::Value(Json::objectValue); target["RemainingAncestor"]["Path"] = GetBasePath(type, uuid); target["RemainingAncestor"]["Type"] = EnumerationToString(type); target["RemainingAncestor"]["ID"] = uuid; } else { target["RemainingAncestor"] = Json::nullValue; } t.Commit(0); return true; } void ServerIndex::FlushThread(ServerIndex* that) { unsigned int sleep; try { std::string sleepString = that->db_->GetGlobalProperty(GlobalProperty_FlushSleep); sleep = boost::lexical_cast<unsigned int>(sleepString); } catch (boost::bad_lexical_cast&) { // By default, wait for 10 seconds before flushing sleep = 10; } LOG(INFO) << "Starting the database flushing thread (sleep = " << sleep << ")"; unsigned int count = 0; while (!that->done_) { boost::this_thread::sleep(boost::posix_time::seconds(1)); count++; if (count < sleep) { continue; } boost::mutex::scoped_lock lock(that->mutex_); that->db_->FlushToDisk(); count = 0; } LOG(INFO) << "Stopping the database flushing thread"; } static void ComputeExpectedNumberOfInstances(DatabaseWrapper& db, int64_t series, const DicomMap& dicomSummary) { const DicomValue* value; const DicomValue* value2; try { if ((value = dicomSummary.TestAndGetValue(DICOM_TAG_IMAGES_IN_ACQUISITION)) != NULL && (value2 = dicomSummary.TestAndGetValue(DICOM_TAG_NUMBER_OF_TEMPORAL_POSITIONS)) != NULL) { // Patch for series with temporal positions thanks to Will Ryder int64_t imagesInAcquisition = boost::lexical_cast<int64_t>(value->AsString()); int64_t countTemporalPositions = boost::lexical_cast<int64_t>(value2->AsString()); std::string expected = boost::lexical_cast<std::string>(imagesInAcquisition * countTemporalPositions); db.SetMetadata(series, MetadataType_Series_ExpectedNumberOfInstances, expected); } else if ((value = dicomSummary.TestAndGetValue(DICOM_TAG_NUMBER_OF_SLICES)) != NULL && (value2 = dicomSummary.TestAndGetValue(DICOM_TAG_NUMBER_OF_TIME_SLICES)) != NULL) { // Support of Cardio-PET images int64_t numberOfSlices = boost::lexical_cast<int64_t>(value->AsString()); int64_t numberOfTimeSlices = boost::lexical_cast<int64_t>(value2->AsString()); std::string expected = boost::lexical_cast<std::string>(numberOfSlices * numberOfTimeSlices); db.SetMetadata(series, MetadataType_Series_ExpectedNumberOfInstances, expected); } else if ((value = dicomSummary.TestAndGetValue(DICOM_TAG_CARDIAC_NUMBER_OF_IMAGES)) != NULL) { db.SetMetadata(series, MetadataType_Series_ExpectedNumberOfInstances, value->AsString()); } } catch (boost::bad_lexical_cast) { } } ServerIndex::ServerIndex(ServerContext& context, const std::string& dbPath) : done_(false), maximumStorageSize_(0), maximumPatients_(0) { listener_.reset(new Internals::ServerIndexListener(context)); if (dbPath == ":memory:") { db_.reset(new DatabaseWrapper(*listener_)); } else { boost::filesystem::path p = dbPath; try { boost::filesystem::create_directories(p); } catch (boost::filesystem::filesystem_error) { } db_.reset(new DatabaseWrapper(p.string() + "/index", *listener_)); } currentStorageSize_ = db_->GetTotalCompressedSize(); // Initial recycling if the parameters have changed since the last // execution of Orthanc StandaloneRecycling(); flushThread_ = boost::thread(FlushThread, this); unstableResourcesMonitorThread_ = boost::thread(UnstableResourcesMonitorThread, this); } ServerIndex::~ServerIndex() { done_ = true; if (flushThread_.joinable()) { flushThread_.join(); } if (unstableResourcesMonitorThread_.joinable()) { unstableResourcesMonitorThread_.join(); } } StoreStatus ServerIndex::Store(const DicomMap& dicomSummary, const Attachments& attachments, const std::string& remoteAet) { boost::mutex::scoped_lock lock(mutex_); listener_->Reset(); DicomInstanceHasher hasher(dicomSummary); try { Transaction t(*this); // Do nothing if the instance already exists { ResourceType type; int64_t tmp; if (db_->LookupResource(hasher.HashInstance(), tmp, type)) { assert(type == ResourceType_Instance); return StoreStatus_AlreadyStored; } } // Ensure there is enough room in the storage for the new instance uint64_t instanceSize = 0; for (Attachments::const_iterator it = attachments.begin(); it != attachments.end(); it++) { instanceSize += it->GetCompressedSize(); } Recycle(instanceSize, hasher.HashPatient()); // Create the instance int64_t instance = db_->CreateResource(hasher.HashInstance(), ResourceType_Instance); DicomMap dicom; dicomSummary.ExtractInstanceInformation(dicom); db_->SetMainDicomTags(instance, dicom); // Detect up to which level the patient/study/series/instance // hierarchy must be created int64_t patient = -1, study = -1, series = -1; bool isNewPatient = false; bool isNewStudy = false; bool isNewSeries = false; { ResourceType dummy; if (db_->LookupResource(hasher.HashSeries(), series, dummy)) { assert(dummy == ResourceType_Series); // The patient, the study and the series already exist bool ok = (db_->LookupResource(hasher.HashPatient(), patient, dummy) && db_->LookupResource(hasher.HashStudy(), study, dummy)); assert(ok); } else if (db_->LookupResource(hasher.HashStudy(), study, dummy)) { assert(dummy == ResourceType_Study); // New series: The patient and the study already exist isNewSeries = true; bool ok = db_->LookupResource(hasher.HashPatient(), patient, dummy); assert(ok); } else if (db_->LookupResource(hasher.HashPatient(), patient, dummy)) { assert(dummy == ResourceType_Patient); // New study and series: The patient already exist isNewStudy = true; isNewSeries = true; } else { // New patient, study and series: Nothing exists isNewPatient = true; isNewStudy = true; isNewSeries = true; } } // Create the series if needed if (isNewSeries) { series = db_->CreateResource(hasher.HashSeries(), ResourceType_Series); dicomSummary.ExtractSeriesInformation(dicom); db_->SetMainDicomTags(series, dicom); } // Create the study if needed if (isNewStudy) { study = db_->CreateResource(hasher.HashStudy(), ResourceType_Study); dicomSummary.ExtractStudyInformation(dicom); db_->SetMainDicomTags(study, dicom); } // Create the patient if needed if (isNewPatient) { patient = db_->CreateResource(hasher.HashPatient(), ResourceType_Patient); dicomSummary.ExtractPatientInformation(dicom); db_->SetMainDicomTags(patient, dicom); } // Create the parent-to-child links db_->AttachChild(series, instance); if (isNewSeries) { db_->AttachChild(study, series); } if (isNewStudy) { db_->AttachChild(patient, study); } // Sanity checks assert(patient != -1); assert(study != -1); assert(series != -1); assert(instance != -1); // Attach the files to the newly created instance for (Attachments::const_iterator it = attachments.begin(); it != attachments.end(); it++) { db_->AddAttachment(instance, *it); } // Attach the metadata std::string now = Toolbox::GetNowIsoString(); db_->SetMetadata(instance, MetadataType_Instance_ReceptionDate, now); db_->SetMetadata(series, MetadataType_LastUpdate, now); db_->SetMetadata(study, MetadataType_LastUpdate, now); db_->SetMetadata(patient, MetadataType_LastUpdate, now); db_->SetMetadata(instance, MetadataType_Instance_RemoteAet, remoteAet); const DicomValue* value; if ((value = dicomSummary.TestAndGetValue(DICOM_TAG_INSTANCE_NUMBER)) != NULL || (value = dicomSummary.TestAndGetValue(DICOM_TAG_IMAGE_INDEX)) != NULL) { db_->SetMetadata(instance, MetadataType_Instance_IndexInSeries, value->AsString()); } if (isNewSeries) { ComputeExpectedNumberOfInstances(*db_, series, dicomSummary); } // Check whether the series of this new instance is now completed SeriesStatus seriesStatus = GetSeriesStatus(series); if (seriesStatus == SeriesStatus_Complete) { db_->LogChange(ChangeType_CompletedSeries, series, ResourceType_Series); } // Mark the parent resources of this instance as unstable MarkAsUnstable(patient, ResourceType_Patient); MarkAsUnstable(study, ResourceType_Study); MarkAsUnstable(series, ResourceType_Series); t.Commit(instanceSize); return StoreStatus_Success; } catch (OrthancException& e) { LOG(ERROR) << "EXCEPTION [" << e.What() << "]" << " (SQLite status: " << db_->GetErrorMessage() << ")"; } return StoreStatus_Failure; } void ServerIndex::ComputeStatistics(Json::Value& target) { boost::mutex::scoped_lock lock(mutex_); target = Json::objectValue; uint64_t cs = currentStorageSize_; assert(cs == db_->GetTotalCompressedSize()); uint64_t us = db_->GetTotalUncompressedSize(); target["TotalDiskSize"] = boost::lexical_cast<std::string>(cs); target["TotalUncompressedSize"] = boost::lexical_cast<std::string>(us); target["TotalDiskSizeMB"] = boost::lexical_cast<unsigned int>(cs / MEGA_BYTES); target["TotalUncompressedSizeMB"] = boost::lexical_cast<unsigned int>(us / MEGA_BYTES); target["CountPatients"] = static_cast<unsigned int>(db_->GetResourceCount(ResourceType_Patient)); target["CountStudies"] = static_cast<unsigned int>(db_->GetResourceCount(ResourceType_Study)); target["CountSeries"] = static_cast<unsigned int>(db_->GetResourceCount(ResourceType_Series)); target["CountInstances"] = static_cast<unsigned int>(db_->GetResourceCount(ResourceType_Instance)); } SeriesStatus ServerIndex::GetSeriesStatus(int64_t id) { // Get the expected number of instances in this series (from the metadata) std::string s = db_->GetMetadata(id, MetadataType_Series_ExpectedNumberOfInstances); size_t expected; try { expected = boost::lexical_cast<size_t>(s); if (expected < 0) { return SeriesStatus_Unknown; } } catch (boost::bad_lexical_cast&) { return SeriesStatus_Unknown; } // Loop over the instances of this series std::list<int64_t> children; db_->GetChildrenInternalId(children, id); std::set<size_t> instances; for (std::list<int64_t>::const_iterator it = children.begin(); it != children.end(); it++) { // Get the index of this instance in the series s = db_->GetMetadata(*it, MetadataType_Instance_IndexInSeries); size_t index; try { index = boost::lexical_cast<size_t>(s); } catch (boost::bad_lexical_cast&) { return SeriesStatus_Unknown; } if (index <= 0 || index > expected) { // Out-of-range instance index return SeriesStatus_Inconsistent; } if (instances.find(index) != instances.end()) { // Twice the same instance index return SeriesStatus_Inconsistent; } instances.insert(index); } if (instances.size() == expected) { return SeriesStatus_Complete; } else { return SeriesStatus_Missing; } } void ServerIndex::MainDicomTagsToJson(Json::Value& target, int64_t resourceId) { DicomMap tags; db_->GetMainDicomTags(tags, resourceId); target["MainDicomTags"] = Json::objectValue; FromDcmtkBridge::ToJson(target["MainDicomTags"], tags); } bool ServerIndex::LookupResource(Json::Value& result, const std::string& publicId, ResourceType expectedType) { result = Json::objectValue; boost::mutex::scoped_lock lock(mutex_); // Lookup for the requested resource int64_t id; ResourceType type; if (!db_->LookupResource(publicId, id, type) || type != expectedType) { return false; } // Find the parent resource (if it exists) if (type != ResourceType_Patient) { int64_t parentId; if (!db_->LookupParent(parentId, id)) { throw OrthancException(ErrorCode_InternalError); } std::string parent = db_->GetPublicId(parentId); switch (type) { case ResourceType_Study: result["ParentPatient"] = parent; break; case ResourceType_Series: result["ParentStudy"] = parent; break; case ResourceType_Instance: result["ParentSeries"] = parent; break; default: throw OrthancException(ErrorCode_InternalError); } } // List the children resources std::list<std::string> children; db_->GetChildrenPublicId(children, id); if (type != ResourceType_Instance) { Json::Value c = Json::arrayValue; for (std::list<std::string>::const_iterator it = children.begin(); it != children.end(); it++) { c.append(*it); } switch (type) { case ResourceType_Patient: result["Studies"] = c; break; case ResourceType_Study: result["Series"] = c; break; case ResourceType_Series: result["Instances"] = c; break; default: throw OrthancException(ErrorCode_InternalError); } } // Set the resource type switch (type) { case ResourceType_Patient: result["Type"] = "Patient"; break; case ResourceType_Study: result["Type"] = "Study"; break; case ResourceType_Series: { result["Type"] = "Series"; result["Status"] = EnumerationToString(GetSeriesStatus(id)); int i; if (db_->GetMetadataAsInteger(i, id, MetadataType_Series_ExpectedNumberOfInstances)) result["ExpectedNumberOfInstances"] = i; else result["ExpectedNumberOfInstances"] = Json::nullValue; break; } case ResourceType_Instance: { result["Type"] = "Instance"; FileInfo attachment; if (!db_->LookupAttachment(attachment, id, FileContentType_Dicom)) { throw OrthancException(ErrorCode_InternalError); } result["FileSize"] = static_cast<unsigned int>(attachment.GetUncompressedSize()); result["FileUuid"] = attachment.GetUuid(); int i; if (db_->GetMetadataAsInteger(i, id, MetadataType_Instance_IndexInSeries)) result["IndexInSeries"] = i; else result["IndexInSeries"] = Json::nullValue; break; } default: throw OrthancException(ErrorCode_InternalError); } // Record the remaining information result["ID"] = publicId; MainDicomTagsToJson(result, id); std::string tmp; tmp = db_->GetMetadata(id, MetadataType_AnonymizedFrom); if (tmp.size() != 0) result["AnonymizedFrom"] = tmp; tmp = db_->GetMetadata(id, MetadataType_ModifiedFrom); if (tmp.size() != 0) result["ModifiedFrom"] = tmp; if (type == ResourceType_Patient || type == ResourceType_Study || type == ResourceType_Series) { result["IsStable"] = !unstableResources_.Contains(id); } return true; } bool ServerIndex::LookupAttachment(FileInfo& attachment, const std::string& instanceUuid, FileContentType contentType) { boost::mutex::scoped_lock lock(mutex_); int64_t id; ResourceType type; if (!db_->LookupResource(instanceUuid, id, type) || type != ResourceType_Instance) { throw OrthancException(ErrorCode_InternalError); } if (db_->LookupAttachment(attachment, id, contentType)) { assert(attachment.GetContentType() == contentType); return true; } else { return false; } } void ServerIndex::GetAllUuids(Json::Value& target, ResourceType resourceType) { boost::mutex::scoped_lock lock(mutex_); db_->GetAllPublicIds(target, resourceType); } bool ServerIndex::GetChanges(Json::Value& target, int64_t since, unsigned int maxResults) { boost::mutex::scoped_lock lock(mutex_); db_->GetChanges(target, since, maxResults); return true; } bool ServerIndex::GetLastChange(Json::Value& target) { boost::mutex::scoped_lock lock(mutex_); db_->GetLastChange(target); return true; } void ServerIndex::LogExportedResource(const std::string& publicId, const std::string& remoteModality) { boost::mutex::scoped_lock lock(mutex_); int64_t id; ResourceType type; if (!db_->LookupResource(publicId, id, type)) { throw OrthancException(ErrorCode_InternalError); } std::string patientId; std::string studyInstanceUid; std::string seriesInstanceUid; std::string sopInstanceUid; int64_t currentId = id; ResourceType currentType = type; // Iteratively go up inside the patient/study/series/instance hierarchy bool done = false; while (!done) { DicomMap map; db_->GetMainDicomTags(map, currentId); switch (currentType) { case ResourceType_Patient: patientId = map.GetValue(DICOM_TAG_PATIENT_ID).AsString(); done = true; break; case ResourceType_Study: studyInstanceUid = map.GetValue(DICOM_TAG_STUDY_INSTANCE_UID).AsString(); currentType = ResourceType_Patient; break; case ResourceType_Series: seriesInstanceUid = map.GetValue(DICOM_TAG_SERIES_INSTANCE_UID).AsString(); currentType = ResourceType_Study; break; case ResourceType_Instance: sopInstanceUid = map.GetValue(DICOM_TAG_SOP_INSTANCE_UID).AsString(); currentType = ResourceType_Series; break; default: throw OrthancException(ErrorCode_InternalError); } // If we have not reached the Patient level, find the parent of // the current resource if (!done) { bool ok = db_->LookupParent(currentId, currentId); assert(ok); } } // No need for a SQLite::Transaction here, as we only insert 1 record db_->LogExportedResource(type, publicId, remoteModality, patientId, studyInstanceUid, seriesInstanceUid, sopInstanceUid); } bool ServerIndex::GetExportedResources(Json::Value& target, int64_t since, unsigned int maxResults) { boost::mutex::scoped_lock lock(mutex_); db_->GetExportedResources(target, since, maxResults); return true; } bool ServerIndex::GetLastExportedResource(Json::Value& target) { boost::mutex::scoped_lock lock(mutex_); db_->GetLastExportedResource(target); return true; } bool ServerIndex::IsRecyclingNeeded(uint64_t instanceSize) { if (maximumStorageSize_ != 0) { uint64_t currentSize = currentStorageSize_ - listener_->GetSizeOfFilesToRemove(); assert(db_->GetTotalCompressedSize() == currentSize); if (currentSize + instanceSize > maximumStorageSize_) { return true; } } if (maximumPatients_ != 0) { uint64_t patientCount = db_->GetResourceCount(ResourceType_Patient); if (patientCount > maximumPatients_) { return true; } } return false; } void ServerIndex::Recycle(uint64_t instanceSize, const std::string& newPatientId) { if (!IsRecyclingNeeded(instanceSize)) { return; } // Check whether other DICOM instances from this patient are // already stored int64_t patientToAvoid; ResourceType type; bool hasPatientToAvoid = db_->LookupResource(newPatientId, patientToAvoid, type); if (hasPatientToAvoid && type != ResourceType_Patient) { throw OrthancException(ErrorCode_InternalError); } // Iteratively select patient to remove until there is enough // space in the DICOM store int64_t patientToRecycle; while (true) { // If other instances of this patient are already in the store, // we must avoid to recycle them bool ok = hasPatientToAvoid ? db_->SelectPatientToRecycle(patientToRecycle, patientToAvoid) : db_->SelectPatientToRecycle(patientToRecycle); if (!ok) { throw OrthancException(ErrorCode_FullStorage); } LOG(INFO) << "Recycling one patient"; db_->DeleteResource(patientToRecycle); if (!IsRecyclingNeeded(instanceSize)) { // OK, we're done break; } } } void ServerIndex::SetMaximumPatientCount(unsigned int count) { boost::mutex::scoped_lock lock(mutex_); maximumPatients_ = count; if (count == 0) { LOG(WARNING) << "No limit on the number of stored patients"; } else { LOG(WARNING) << "At most " << count << " patients will be stored"; } StandaloneRecycling(); } void ServerIndex::SetMaximumStorageSize(uint64_t size) { boost::mutex::scoped_lock lock(mutex_); maximumStorageSize_ = size; if (size == 0) { LOG(WARNING) << "No limit on the size of the storage area"; } else { LOG(WARNING) << "At most " << (size / MEGA_BYTES) << "MB will be used for the storage area"; } StandaloneRecycling(); } void ServerIndex::StandaloneRecycling() { // WARNING: No mutex here, do not include this as a public method Transaction t(*this); Recycle(0, ""); t.Commit(0); } bool ServerIndex::IsProtectedPatient(const std::string& publicId) { boost::mutex::scoped_lock lock(mutex_); // Lookup for the requested resource int64_t id; ResourceType type; if (!db_->LookupResource(publicId, id, type) || type != ResourceType_Patient) { throw OrthancException(ErrorCode_ParameterOutOfRange); } return db_->IsProtectedPatient(id); } void ServerIndex::SetProtectedPatient(const std::string& publicId, bool isProtected) { boost::mutex::scoped_lock lock(mutex_); // Lookup for the requested resource int64_t id; ResourceType type; if (!db_->LookupResource(publicId, id, type) || type != ResourceType_Patient) { throw OrthancException(ErrorCode_ParameterOutOfRange); } // No need for a SQLite::Transaction here, as we only make 1 write to the DB db_->SetProtectedPatient(id, isProtected); if (isProtected) LOG(INFO) << "Patient " << publicId << " has been protected"; else LOG(INFO) << "Patient " << publicId << " has been unprotected"; } void ServerIndex::GetChildInstances(std::list<std::string>& result, const std::string& publicId) { result.clear(); boost::mutex::scoped_lock lock(mutex_); ResourceType type; int64_t top; if (!db_->LookupResource(publicId, top, type)) { throw OrthancException(ErrorCode_UnknownResource); } if (type == ResourceType_Instance) { // The resource is already an instance: Do not go down the hierarchy result.push_back(publicId); return; } std::stack<int64_t> toExplore; toExplore.push(top); std::list<int64_t> tmp; while (!toExplore.empty()) { // Get the internal ID of the current resource int64_t resource = toExplore.top(); toExplore.pop(); if (db_->GetResourceType(resource) == ResourceType_Instance) { result.push_back(db_->GetPublicId(resource)); } else { // Tag all the children of this resource as to be explored db_->GetChildrenInternalId(tmp, resource); for (std::list<int64_t>::const_iterator it = tmp.begin(); it != tmp.end(); it++) { toExplore.push(*it); } } } } void ServerIndex::SetMetadata(const std::string& publicId, MetadataType type, const std::string& value) { boost::mutex::scoped_lock lock(mutex_); ResourceType rtype; int64_t id; if (!db_->LookupResource(publicId, id, rtype)) { throw OrthancException(ErrorCode_UnknownResource); } db_->SetMetadata(id, type, value); } void ServerIndex::DeleteMetadata(const std::string& publicId, MetadataType type) { boost::mutex::scoped_lock lock(mutex_); ResourceType rtype; int64_t id; if (!db_->LookupResource(publicId, id, rtype)) { throw OrthancException(ErrorCode_UnknownResource); } db_->DeleteMetadata(id, type); } bool ServerIndex::LookupMetadata(std::string& target, const std::string& publicId, MetadataType type) { boost::mutex::scoped_lock lock(mutex_); ResourceType rtype; int64_t id; if (!db_->LookupResource(publicId, id, rtype)) { throw OrthancException(ErrorCode_UnknownResource); } return db_->LookupMetadata(target, id, type); } bool ServerIndex::ListAvailableMetadata(std::list<MetadataType>& target, const std::string& publicId) { boost::mutex::scoped_lock lock(mutex_); ResourceType rtype; int64_t id; if (!db_->LookupResource(publicId, id, rtype)) { throw OrthancException(ErrorCode_UnknownResource); } return db_->ListAvailableMetadata(target, id); } bool ServerIndex::LookupParent(std::string& target, const std::string& publicId) { boost::mutex::scoped_lock lock(mutex_); ResourceType type; int64_t id; if (!db_->LookupResource(publicId, id, type)) { throw OrthancException(ErrorCode_UnknownResource); } int64_t parentId; if (db_->LookupParent(parentId, id)) { target = db_->GetPublicId(parentId); return true; } else { return false; } } uint64_t ServerIndex::IncrementGlobalSequence(GlobalProperty sequence) { boost::mutex::scoped_lock lock(mutex_); std::auto_ptr<SQLite::Transaction> transaction(db_->StartTransaction()); transaction->Begin(); uint64_t seq = db_->IncrementGlobalSequence(sequence); transaction->Commit(); return seq; } void ServerIndex::LogChange(ChangeType changeType, const std::string& publicId) { boost::mutex::scoped_lock lock(mutex_); std::auto_ptr<SQLite::Transaction> transaction(db_->StartTransaction()); transaction->Begin(); int64_t id; ResourceType type; if (!db_->LookupResource(publicId, id, type)) { throw OrthancException(ErrorCode_UnknownResource); } db_->LogChange(changeType, id, type); transaction->Commit(); } void ServerIndex::DeleteChanges() { boost::mutex::scoped_lock lock(mutex_); db_->ClearTable("Changes"); } void ServerIndex::DeleteExportedResources() { boost::mutex::scoped_lock lock(mutex_); db_->ClearTable("ExportedResources"); } void ServerIndex::GetStatistics(Json::Value& target, const std::string& publicId) { boost::mutex::scoped_lock lock(mutex_); ResourceType type; int64_t top; if (!db_->LookupResource(publicId, top, type)) { throw OrthancException(ErrorCode_UnknownResource); } std::stack<int64_t> toExplore; toExplore.push(top); int countInstances = 0; int countSeries = 0; int countStudies = 0; uint64_t compressedSize = 0; uint64_t uncompressedSize = 0; while (!toExplore.empty()) { // Get the internal ID of the current resource int64_t resource = toExplore.top(); toExplore.pop(); ResourceType thisType = db_->GetResourceType(resource); if (thisType == ResourceType_Instance) { std::list<FileContentType> f; db_->ListAvailableAttachments(f, resource); for (std::list<FileContentType>::const_iterator it = f.begin(); it != f.end(); it++) { FileInfo attachment; if (db_->LookupAttachment(attachment, resource, *it)) { compressedSize += attachment.GetCompressedSize(); uncompressedSize += attachment.GetUncompressedSize(); } } countInstances++; } else { switch (thisType) { case ResourceType_Study: countStudies++; break; case ResourceType_Series: countSeries++; break; default: break; } // Tag all the children of this resource as to be explored std::list<int64_t> tmp; db_->GetChildrenInternalId(tmp, resource); for (std::list<int64_t>::const_iterator it = tmp.begin(); it != tmp.end(); it++) { toExplore.push(*it); } } } target = Json::objectValue; target["DiskSize"] = boost::lexical_cast<std::string>(compressedSize); target["DiskSizeMB"] = boost::lexical_cast<unsigned int>(compressedSize / MEGA_BYTES); target["UncompressedSize"] = boost::lexical_cast<std::string>(uncompressedSize); target["UncompressedSizeMB"] = boost::lexical_cast<unsigned int>(uncompressedSize / MEGA_BYTES); switch (type) { // Do NOT add "break" below this point! case ResourceType_Patient: target["CountStudies"] = countStudies; case ResourceType_Study: target["CountSeries"] = countSeries; case ResourceType_Series: target["CountInstances"] = countInstances; case ResourceType_Instance: default: break; } } void ServerIndex::UnstableResourcesMonitorThread(ServerIndex* that) { int stableAge = GetGlobalIntegerParameter("StableAge", 60); if (stableAge <= 0) { stableAge = 60; } LOG(INFO) << "Starting the monitor for stable resources (stable age = " << stableAge << ")"; while (!that->done_) { // Check for stable resources each second boost::this_thread::sleep(boost::posix_time::seconds(1)); boost::mutex::scoped_lock lock(that->mutex_); while (!that->unstableResources_.IsEmpty() && that->unstableResources_.GetOldestPayload().GetAge() > static_cast<unsigned int>(stableAge)) { // This DICOM resource has not received any new instance for // some time. It can be considered as stable. UnstableResourcePayload payload; int64_t id = that->unstableResources_.RemoveOldest(payload); // Ensure that the resource is still existing before logging the change if (that->db_->IsExistingResource(id)) { switch (payload.type_) { case Orthanc::ResourceType_Patient: that->db_->LogChange(ChangeType_StablePatient, id, ResourceType_Patient); break; case Orthanc::ResourceType_Study: that->db_->LogChange(ChangeType_StableStudy, id, ResourceType_Study); break; case Orthanc::ResourceType_Series: that->db_->LogChange(ChangeType_StableSeries, id, ResourceType_Series); break; default: throw OrthancException(ErrorCode_InternalError); } //LOG(INFO) << "Stable resource: " << EnumerationToString(payload.type_) << " " << id; } } } LOG(INFO) << "Closing the monitor thread for stable resources"; } void ServerIndex::MarkAsUnstable(int64_t id, Orthanc::ResourceType type) { // WARNING: Before calling this method, "mutex_" must be locked. assert(type == Orthanc::ResourceType_Patient || type == Orthanc::ResourceType_Study || type == Orthanc::ResourceType_Series); unstableResources_.AddOrMakeMostRecent(id, type); //LOG(INFO) << "Unstable resource: " << EnumerationToString(type) << " " << id; } void ServerIndex::LookupTagValue(std::list<std::string>& result, DicomTag tag, const std::string& value) { result.clear(); boost::mutex::scoped_lock lock(mutex_); std::list<int64_t> id; db_->LookupTagValue(id, tag, value); for (std::list<int64_t>::const_iterator it = id.begin(); it != id.end(); it++) { result.push_back(db_->GetPublicId(*it)); } } void ServerIndex::LookupTagValue(std::list<std::string>& result, const std::string& value) { result.clear(); boost::mutex::scoped_lock lock(mutex_); std::list<int64_t> id; db_->LookupTagValue(id, value); for (std::list<int64_t>::const_iterator it = id.begin(); it != id.end(); it++) { result.push_back(db_->GetPublicId(*it)); } } }